Queries result takes time on CDC SQL database

I have a DB. CDC is applied almost all tables(50+) of DB.
Is CDC data slow down the DML queries result like select query?
I want to ask this question because my DML queries takes time now.
Or is there any other reason of slowing down DB?

It is true that CDC will have an impact on the overall performance and the same can be addressed by following standard guidelines presented by Microsoft SQL Server. For example, Large number VLFs in Transaction Log may also cause slowness since log reader
will have to traverse too many fragmented portions of T-Log to capture the changed records.
Likewise there are many other factors, Refer the following link for more details including Tuning opportunities:
https://technet.microsoft.com/en-us/library/dd266396%28v=sql.100%29.aspx?f=255&MSPPError=-2147217396
Good Luck! Please Mark This As Answer if it solved your issue. Please Vote This As Helpful if it helps to solve your issue

Similar Messages

  • UUT Result log in SQL database?

    Hello All
    I am facing a requirement where i want to log results of sequence execution[only the complete seq result and NOT step results] to SQL database BUT as and when the UUT_ID varies the entry should be made into a new column with the same name as UUT_ID. As of now whenever i execute a sequence with different UUT_ID, a new row gets added as a result.
    Please reply back with your suggestions asap, as this very urgent for me :-)
    Regards
    Nitin Goel

    But in my opinion, that's just a way to display the results - not a way to store the data. I also have thousands of test cases and tens of thousands of different UUT each month. Each UUT has a separate record (a unique row) and yet if I design a proper query and report format, I can show the results exactly like you have. The design of a relational database should not be the same as a report design, imho.
    A database is designed to have unlimited rows (dependent only on storage space). I'm not sure a db can have unlimited columns. That would be something you could check with your dba. Dynamically adding a new column and doing an update to an existing row would probaby be much slower than just doing an insert.

  • An error has occurred while accessing SQL database or system resources. If this is the first time you have seen this message, please try again later. If this problem persists, please contact your administrator.

    I have SP Server 2010, and when I try to DELETE a rule within an existing Audience, "Property (Account Name) = domain/username", I get this error, "An error has occurred while accessing SQL database or system resources. If this
    is the first time you have seen this message, please try again later. If this
    problem persists, please contact your administrator."  When I try to "MODIFY" the rule I get this error, "One or more values typed on this page are not valid. Check the text for the indicated fields." 
    The last time I checked it was working, I'm not aware of any new updates installed recently?  I did a full Profile Synchronization as well, but still not working, please advise? -- Evenstarline

    Hi Sara,
    First of all thank you very much for your prompt responses. Here are my comments to each of your suggestions below, and just to let you know I am using a Farm Admin account.  I
    was able to do this way after we upgraded from SP 2007 to SP 2010 as well.   I would like to mention I'm not a SP expert, just been given the responsibility due to another person handling it just left, so apologize with some of
    my novice questions below?
    1. When I change the Operators to "Contains" or "Not Contains" get generates this error below.
         Error generating in red towards top of the audience page..."One or more values typed on this page are not valid.  Check the text for the indicated fields."
         Error occurred where you enter your "Value"..."Could not resolve the user identity. Please re-enter the account name."
    2. We have a 3-server-tier topology (SPWeb, SPDB, and SPFarm).  Does the updates only apply to where the Central Admin is installed, which is the "SPFarm"?  I checked all
    3 servers, and NONE of the updates (KB2899494, KB2889845, and KB2883055) you'd mentioned are installed.
    3. I'm new to IISRET, I need to be extra cautious of what I run in production, is this safe to run with no problem?  What does it do?  And How do I run it?
    4. I'm also new to viewing the ULS log.  I'd just downloaded a viewer for it.  I'm assuming the only logs I need to be concern with viewing are within the SPAdmin (where Central
    Admin is installed)?  There's so many of them, what should I be looking for exactly?
    Evenstarline 

  • Execution time, elapsed time  of an sql query

    Can you please tell me how to get the execution time, elapsed time of an sql query

    user8680248 wrote:
    I am running query in the database
    I would like to know how long the query take the time to completeWhy? That answer can be totally meaningless as the VERY SAME query on the VERY SAME data on the VERY SAME database in the VERY SAME Oracle session can and will show DIFFERENT execution times.
    So why do you want to know a specific query's execution time? What do you expect that to tell you?
    If you mean that you want to know how long an existing query being executed is still going to take - that's usually quite difficult to determine. Oracle does provide a view on so-called long operations. However, only certain factors of a query's execution will trigger that this query is a long operation - and only for those specific queries will there be long operation stats that provide an estimated completion time.
    If your slow and long running query does not show in long operation, then Oracle does not consider it a long operation - it fails to meet the specific criteria and factors required as a long operation. This is not a bug or an error. Simply that your query does not meet the basic requirements to be viewed as a long operation.
    Oracle however provides the developer with the means to create long operations (using PL/SQL). You need to know and do the following:
    a) need to know how many units of work to do (e.g. how many fetches/loop iterations/rows your code will process)
    b) need to know how many units of work thus far done
    c) use the DBMS_APPLICATION_INFO package to create a long operation and continually update the operation with the number of work units thus far done
    It is pretty easy to implement this in PL/SQL processing code (assuming requirements a and b can be met) - and provide long operation stats and estimated completion time for the DBA/operators/users of the database, waiting on your process to complete.

  • Concerns over switching between new Azure SQL Database Service Tiers

    Windows Azure's new SQL Database
    service tier pricing model will be put into effect in less than 12 months. We currently have SQL Databases on the Business and Web Edition pricing models.
    We recently asked Azure Support a number of questions around the scalability and ability to switch between these tiers. The responses so far have been far from encouraging:
    Q: If we exceed the criteria for a given tier (see http://msdn.microsoft.com/library/azure/dn741336.aspx), how will Azure respond? For example, if we are on the S1 service tier, and we exceed the maximum number of sessions (200), will any new sessions be
    blocked until we manually increase the service tier? Or will you automatically move (and bill) us to/for the next tier level?
    A: If you exceed the criteria of the existing tier, you shall be notified of performance issue like throttling. Users may experience slowness and blocking. There will not be any automatic upgrade.
    Q: So to confirm, if we suddenly experience increased, unanticipated client activity overnight due to our web site becoming more popular, you will be blocking any user sessions over and above our performance level limit, until we manually make the transition
    to the next level? Is there no “overdraft” facility/scalability in this respect? If this scenario took place within a short space of time, how would we have time to react? Simply blocking new sessions and preventing customers from using our site is not acceptable.
    A: I’m sure you understand that SQL Azure database is a shared resource and hence we cap the resources for individual subscription so that a fair service is provided to all the concurrent users. If you feel that you cannot compromise on the user experience
    then you shall think of an edition that best suits you. Please refer to the following document to best understand the throttling and its impact. http://social.technet.microsoft.com/wiki/contents/articles/1541.windows-azure-sql-database-connection-management.aspx#Throttling_Limits
    Q: Since the changing of the performance level could take several minutes/hours depending on the size of the database, would we experience any downtime/degradation of performance of the database during this period?
    A: You might
    Q: If downtime/degradation of performance are a possibility whilst we switch service tiers, what do Microsoft recommend we do to safeguard against this? Should we create a second database on a higher tier level and then export the data from the lower tier and
    import to the higher, before switching over? How do Microsoft recommend we switch tiers in a production environment with minimal disruption?
    A: If you want to upgrade the tiers due to degradation of performance, you will have to create another database on a higher tier and import the data.
    One of our key reasons for moving to Azure hosting was the seamless scalability it appeared to offered. As you can imagine, the responses above are a major concern for our production environment. Does anyone else have any thoughts or concerns in this respect?

    James,
    I reviewed the support incident you referred to and believe that the questions you asked may have been answered later in the interaction you had with the engineer.  For the benefit of others viewing this forum I want to reiterate the answers
    to your key questions here as well.
    Q. What is the behavior when you reach the limits for the service tier?
    A. Each service tier currently has limits on 4 different dimensions (CPU, physical reads, log writes and memory) of resource consumption.  When you reach one of the limits, the behavior depends on which resource limit you are hitting, but
    generally speaking is consistent with the behavior you would see with a similar hardware limit in the SQL Server box product.  For example, when you reach the CPU limit your queries will start showing more SOS_SCHEDULER_YIELD waits,
    the memory limit will cause a higher percentage of pages to be read from disk instead of the buffer cache (PAGEIOLATCH_xx waits), etc.  This set of limits does not directly abort any of your queries--they just run longer as they
    vie for the fixed set of resources made available to your database.  In a system where the load greatly exceeds the resources, queries may start to time out.
    The new service tiers continue with the Web/Business edition behavior of limiting the number of sessions and concurrent requests (worker threads) you can have.  When you exceed these limits you'll get error 10928.  Note that
    each tier in Basic/Standard/Premium have different values for these limits (http://msdn.microsoft.com/en-us/library/azure/dn369873.aspx) than Web/Business and thus may encounter the errors at different usage levels.  The key is choosing the appropriate
    service level for the application, and to facilitate that the sys.resource_stats view shows historical usage information so you know where you stand as far as reaching any of the limits.  This resource consumption data is also available in the portal.
    Q. Is it possible to switch service tiers, and is there any disruption when you do so?
    A. Yes, you can change between service tiers as described in this MSDN documentation (http://msdn.microsoft.com/en-us/library/azure/dn369872.aspx).  This can be done through the portal, powershell, or REST APIs.  [The preview currently has a restriction
    where legacy servers don't support switching to Basic/Standard which is expected to be removed in the near future.]  The link above outlines the limits on the number of tier changes you can do in a 24 hour period, expected time to perform the change, and
    the client disconnect that occurs.
    If you have further questions, feel free to re-engage on the support incident or reply to this forum thread.

  • Intermittant (not constant) SQL database ShareWebDb login failed events 3760 and 2424

    My Sharepoint 3.0 services that were bundled with SBS 2008 appear to be working fine. Companyweb, etc, all work w/o issue, so I don't want to take violent action, but these errors seem to be accellerating so I would love to know if there is a known
    solution.
    I typically receive the following (these are in reverse time order, latest at the top, just as they appear in the event log):
    Error Event ID: 3760 SQL Database 'ShareWebDb' on SQL Server instance 'np:\\.\pipe\MSSQL$Microsoft##SSEE\sql\query' not found. Additional error information from SQL Server is included below. Cannot open database "ShareWebDb" requested
    by the login. The login failed. Login failed for user 'NT AUTHORITY\NETWORK SERVICE'.
    one second earlier:Error Event ID: 2424 The update cannot be started because the content sources cannot be accessed. Fix the errors and try the update again. Context: Application 'Search', Catalog 'index file
    on the search server Search'
    Information Event ID: 18456 No description, here is the Details:
    <Event xmlns="http://schemas.microsoft.com/win/2004/08/events/event">
    <System>
    <Provider
    Name="MSSQL$MICROSOFT##SSEE" />
    <EventID Qualifiers="49152">18456</EventID>
    <Level>0</Level>
    <Task>4</Task>
    <Keywords>0x90000000000000</Keywords>
    <TimeCreated
    SystemTime="2010-11-12T18:35:06.000Z" />
    <EventRecordID>491461</EventRecordID>
    <Channel>Application</Channel>
    <Computer>myserver.mydomain.local</Computer>
    <Security
    UserID="S-1-5-20" />
    </System>
    <EventData>
    <Data>NT AUTHORITY\NETWORK SERVICE</Data>
    <Data>[CLIENT: <named pipe>]</Data>
    <Binary>184800000E000000170000005000450054003300310030005C004D004900430052004F0053004F00460054002300230053005300450045000000070000006D00610073007400650072000000</Binary>
    </EventData>
    </Event>
    Two seconds earlier
    Information Event ID 17137 No description, here is the Details:
    <Event xmlns="http://schemas.microsoft.com/win/2004/08/events/event">
    <System>
    <Provider
    Name="MSSQL$MICROSOFT##SSEE" />
    <EventID Qualifiers="16384">17137</EventID>
    <Level>4</Level>
    <Task>2</Task>
    <Keywords>0x80000000000000</Keywords>
    <TimeCreated
    SystemTime="2010-11-12T18:35:04.000Z" />
    <EventRecordID>491460</EventRecordID>
    <Channel>Application</Channel>
    <Computer>mysever.mydomain.local</Computer>
    <Security
    UserID="S-1-5-20" />
    </System>
    <EventData>
    <Data>ShareWebDb</Data>
    <Binary>F14200000A000000170000005000450054003300310030005C004D004900430052004F0053004F00460054002300230053005300450045000000070000006D00610073007400650072000000</Binary>
    </EventData>
    </Event>
    I get more 3760 errors than 2424's. However, the
    2424's do seem to usually correspond to a 3760 18456 17137 occurance and the
    3760 are always preceeded by 18456 17137. I was getting one or two sets a day, lately that has increased to 5 or 10.
    I've done a lot of
    web research on Sharepoint Errors 3760 and
    2424 and I see a number of people are experiencing the the same issue. Most of what I've come across for solutions do not apply, since errors are intermittent, not chronic and my public remote URL has already been added
    to BackConnectionHostNames as per:
    http://blogs.technet.com/b/sbs/archive/2009/08/05/sharepoint-services-3-search-event-errors-after-an-sbs-2008-update-rollup.aspx
    I installed the SQL Server Management Studio Express so I could try to track down the permissions on this database. I see the SQL server instance, MYSERVER\Microsoft##ssee, and it shows as running, but when I try to connect to it I get the following
    error: "A network-related or instance-specific error occurred while establishing a connection to SQL Server. The server was not found or was not accessible. Verify that the instance name is correct and that SQL Server is configured to allow remote
    connections. (provider: SQL Network Interfaces, error: 26 - Error Locating Server/Instance Specified) (Microsoft SQL Server, Error: -1)" I'm assuming that is because I don't have access permission for that internal database.
    I don't see any errors in the event log related to failed logins or any errors at all from the SQL server, but there are the time matching informational entries right before the error events and lots of other 17137's and 2803's from MSSQL$Microsoft##ssee

    Thank you Kevin, that does seem to match my symptoms and I have made the change suggested. Just incase that page becomes unreachable by others, here is the part of the page that applies to this:
    Issue:
    =========
    Receive event error 3760 and 6398 on SBS 2008
    Cause:
    ==========
    Issue was caused by incorrect Auto Close option for ShareWebDb database
    Solution:
    ==========
    1. On the SBS 2008 server, click Start -> All Programs -> Microsoft SQL Server 2005, right click SQL Server Management Studio Express and choose run as administrator.
    NOTE: You must run it as administrator to perform steps below.
    2. In the Server Name box, type \\.\pipe\MSSQL$Microsoft##SSEE\sql\query, click connect.
    3. Expand Database node and right click the ShareWebDb then choose properties.
    4. Click Options in the left pane, in the right pane, please check whether the Auto Close option is set to True, if so, please change it to False.

  • Merge to Insert or update records in SQL Database

    Hello ,
    I am having hard time with creating a Stored Procedure to insert/update the selected records from Oracle data base to SQL database using BizTalk.
    ALTER PROCEDURE [dbo].[uspInsertorUpdateINF]
    @dp_id char(32),
    @dv_id char(32),
    @em_number char(12),
    @email varchar(50),
    @emergency_relation char(32),
    @option1 char(16),
    @status char(20),
    @em_id char(35),
    @em_title varchar(64),
    @date_hired datetime
    AS
    BEGIN
    SET NOCOUNT ON;
    MERGE [dbo].[em] AS [Targ]
    USING (VALUES (@dp_id, @dv_id , @em_number, @email, @emergency_relation, @option1, @status, @em_id, @em_title, @date_hired))
    AS [Sourc] (dp_id, dv_id, em_number, email, emergency_relation, option1, status, em_id, em_title, date_hired)
    ON [Targ].em_id = [Sourc].em_id
    WHEN MATCHED THEN
    UPDATE SET dp_id = [Sourc].dp_id,
    dv_id = [Sourc].dv_id,
    em_number = [Sourc].em_number,
    email = [Sourc].email,
    emergency_relation = [Sourc].emergency_relation,
    option1 = [Sourc].option1,
    status = [Sourc].status,
    em_title = [Sourc].em_title,
    date_hired = [Sourc].date_hired
    WHEN NOT MATCHED BY TARGET THEN
    INSERT (dp_id, dv_id, em_number, email, emergency_relation, option1, status, em_id, em_title,date_hired)
    VALUES ([Sourc].dp_id, [Sourc].dv_id, [Sourc].em_number, [Sourc].email, [Sourc].emergency_relation, [Sourc].option1, [Sourc].status, [Sourc].em_id, [Sourc].em_title, [Sourc].date_hired);
    END;
    I am getting an error like
    WcfSendPort_SqlAdapterBinding_
    TypedProcedures_dbo_Custom" with URL "mssql://abc//def?". It will be retransmitted after the retry interval specified for this Send Port. Details:"System.Data.SqlClient.SqlException (0x80131904): Procedure or function 'uspInsertorUpdateINF'
    expects parameter '@dp_id', which was not supplied
    I cannot give the Oracle Database name directly in the stored Procedure
    I am stuck with this, and cannot figure out since I am new to SQL Queries and Stored Procedures. Any help is greatly appreciated.
    Thanks

    Hi sid_siv,
    Only the first record is inserted because of the scalar variables of the stored procedure(SP), when you call the SP, only one row is passed. To get all rows inserted, you can either declare a
    table-valued parameter
    (TVP) for the SP or using an Oracle linked server in the SP.
    Create a stored procedure with a table-valued parameter
    As you mentioned linked server is not good in your case, then TVP SP can be the only option. Regarding how to call a SP with TVP parameter in BizTalk, that's a typically BizTalk question then. I would suggest you post your question in a dedicated
    BizTalk Server forum. It is more appropriate and more experts will assist you.
    If you have any question, feel free to let me know.
    Eric Zhang
    TechNet Community Support

  • How can I insert an image(JPG) into a SQL database?

    I have written an application to perform some tests in the lab.  After gathering the data I store it into a SQL Server.  I successfully managed to export all my raw numbers and calculations to the DB.  However I have not been able to export a screenshot of the Front panel.  I can write the screenshot to a JPG file just fine.  The problem is putting the image (binary data) into the DB.
    So from what work I've managed to do, two questions come to mind:
    1)  Is there a way to compress the screenshot, using standard JPG compression, without actually writing out to a file.  ie:  Just have a binary stream of the compressed image that I can write to the SQL database.
    2)  Even if I don't use compression, I can't write any kind of data that appears as an image.  What kind of data do I need to be writing from LabView to store it properly in an 'image' or 'binary' type field in a SQL database?  I am not looking for answer "binary data."  Nor do I want the answer that is stated in the Help file for the DB Connectivity Toolkit(type - Data Cluster).  What I need is to know what kind of conversions I need to do on the image data that comes out of the FP.GetScaledImage.
    I am using LabView 8.0 and have the Database connectivity toolkit.  We are running SQL Server 2000 and 2005.

    I also hit this limit.  In the end I just wrote the JPGs to disk, and stored the file path in the DB.  Not the ideal solution, I know.
    I'm trying to remember now, exactly what the problem was.  You're right to say the image data type should handle up to 2GB.  I think in the end, it was an issue of the LabView toolkit not playing well with newer versions of SQL Server, which have the image type.  Re-reading the error though, it sounds like it might be an issue with the ODBC driver.  FYI, 8KB used to be the max size for the largest types in SQL Server (I think binary in 2000).  I suspect this limit was coded into the driver or LabView, and can't handle the newer, larger types.
    That being said, the LabView DB toolkit was either written poorly or written a long time ago.  It has issues with generating SQL queries that have the proper quotes.  For example, if your column names have spaces in them, the toolkit will be unable to generate a valid query, without some modification of the toolkit *.vi's.  Images in a database is not a new concept and not new to me.  I've never used a language where it was more complicated or difficult to get working.  Unless you absolutely have to, I'd work out a way to just store the JPGs to disk, with the path in the DB.  Spending too much time on this defeats the purpose of using LabView.
    If NI decides to update the DB toolkit, then it might be worth giving it a try again.
    In case any NI people are reading this, this is a not so subtle hint that you need to update the DB toolkit to support modern DB features.  Especially, if NI is charging $1000 for it.

  • How to find out the execution time of a sql inside a function

    Hi All,
    I am writing one function. There is only one IN parameter. In that parameter, i will pass one SQL select statement. And I want the function to return the exact execution time of that SQL statement.
    CREATE OR REPLACE FUNCTION function_name (p_sql IN VARCHAR2)
    RETURN NUMBER
    IS
    exec_time NUMBER;
    BEGIN
    --Calculate the execution time for the incoming sql statement.
    RETURN exec_time;
    END function_name;
    /

    Please note that wrapping query in a "SELECT COUNT(*) FROM (<query>)" doesn't necessarily reflect the execution time of the stand-alone query because the optimizer is smart and might choose a completely different execution plan for that query.
    A simple test case shows the potential difference of work performed by the database:
    Oracle Database 10g Express Edition Release 10.2.0.1.0 - Production
    Session altered.
    SQL>
    SQL> drop table count_test purge;
    Table dropped.
    Elapsed: 00:00:00.17
    SQL>
    SQL> create table count_test as select * from all_objects;
    Table created.
    Elapsed: 00:00:02.56
    SQL>
    SQL> alter table count_test add constraint pk_count_test primary key (object_id)
    Table altered.
    Elapsed: 00:00:00.04
    SQL>
    SQL> exec dbms_stats.gather_table_stats(ownname=>null, tabname=>'COUNT_TEST')
    PL/SQL procedure successfully completed.
    Elapsed: 00:00:00.29
    SQL>
    SQL> set autotrace traceonly
    SQL>
    SQL> select * from count_test;
    5326 rows selected.
    Elapsed: 00:00:00.10
    Execution Plan
    Plan hash value: 3690877688
    | Id  | Operation         | Name       | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT  |            |  5326 |   431K|    23   (5)| 00:00:01 |
    |   1 |  TABLE ACCESS FULL| COUNT_TEST |  5326 |   431K|    23   (5)| 00:00:01 |
    Statistics
              1  recursive calls
              0  db block gets
            419  consistent gets
              0  physical reads
              0  redo size
         242637  bytes sent via SQL*Net to client
           4285  bytes received via SQL*Net from client
            357  SQL*Net roundtrips to/from client
              0  sorts (memory)
              0  sorts (disk)
           5326  rows processed
    SQL>
    SQL> select count(*) from (select * from count_test);
    Elapsed: 00:00:00.00
    Execution Plan
    Plan hash value: 572193338
    | Id  | Operation             | Name          | Rows  | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT      |               |     1 |     5   (0)| 00:00:01 |
    |   1 |  SORT AGGREGATE       |               |     1 |            |          |
    |   2 |   INDEX FAST FULL SCAN| PK_COUNT_TEST |  5326 |     5   (0)| 00:00:01 |
    Statistics
              1  recursive calls
              0  db block gets
             16  consistent gets
              0  physical reads
              0  redo size
            412  bytes sent via SQL*Net to client
            380  bytes received via SQL*Net from client
              2  SQL*Net roundtrips to/from client
              0  sorts (memory)
              0  sorts (disk)
              1  rows processed
    SQL>As you can see the number of blocks processed (consistent gets) is quite different. You need to actually fetch all records, e.g. using a PL/SQL block on the server to find out how long it takes to process the query, but that's not that easy if you want to have an arbitrary query string as input.
    Regards,
    Randolf
    Oracle related stuff blog:
    http://oracle-randolf.blogspot.com/
    SQLTools++ for Oracle:
    http://www.sqltools-plusplus.org:7676/
    http://sourceforge.net/projects/sqlt-pp/

  • Oracle text related internal procedure taking a lot of time in our Production database

    Hi,
    I am on Oracle 11.2.0.2 on Linux. I have Oracle Text implemented in all my databases for fuzzy search. I am seeing the following Oracle TExt specific internal procedure to be among the Top SQL in my AWR in production. This is during business time.
    SQL ordered by Elapsed Time
    Resources reported for PL/SQL code includes the resources used by
    all SQL statements called by the code.
    % Total DB Time is the Elapsed Time of the SQL statement divided
    into the Total Database Time multiplied by 100
    %Total - Elapsed Time as a percentage of Total DB time
    %CPU - CPU Time as a percentage of Elapsed Time
    %IO - User I/O Time as a percentage of Elapsed Time
    Captured SQL account for 59.3% of Total DB Time (s): 120,379
    Captured PL/SQL account for 33.8% of Total DB Time (s): 120,379
    Elapsed Time (s)
    Executions
    Elapsed Time per Exec (s)
    %Total
    %CPU
    %IO
    SQL Id
    SQL Module
    SQL Text
    23,476.22
    205,095
    0.11
    19.50
    16.21
    7.88
    ddr8uck5s5kp3
    begin ctxsys.drvdml.com_sync_i...
    Note that the sql id ddr8uck5s5kp3 has this sql:
    begin ctxsys.drvdml.com_sync_index(:idxname, :idxmem, :partname); end;
    Also note that I have the procedure to optimize the indexes (ctx_ddl.optimize_index in FULL mode) set up every night to run at 3 am for all our Oracle Text indexes.  Is there anything else needed. I don't know why the procedure I showed above in the AWR report takes so much time and why it is among our Top sql.
    I will be very thankful for guidance in this regard.
    Thanks,
    OrauserN

    This is the internal call which drives the SYNC call for a text index. Effectively all the indexing of new and updated data in your text-indexed table is contained within this call.
    If you're using parallel SYNC you will see this call contained with a SELECT query - that select is executed as a parallel query on a table function, which is the way we divide up the work between parallel slaves.

  • Insuffiecient SQL database Permission for User

    Hi All,
    The following error is coming in the Event Viewer:
    Insufficient SQL database permissions for user 'Name: NNOK\SPPEAPO_SVC001_ES SID: S-1-5-21-2610070952-2089559051-1579118431-1720 ImpersonationLevel: Impersonation' in database 'SharePoint_AdminContent_6fa0b459-ed7c-481f-85c4-b81753e39346' on SQL Server instance
    'NOKWDCFISP'. 
    The EXECUTE permission was denied on the object 'proc_GetTimerRunningJobs', database 'SharePoint_Config', schema 'dbo'.
    Kindly reply if you know the solution for this.
    Thanks in Advance.
    Pravin Dhulap

    Hi Pravin,
    You can grant "WSS_Content_Application_Pools" execute permission to the procedure "proc_GetTimerRunningJobs"
     under SharePoint _Config database, then check results again. 
    Here is another post you can take a look.
    https://social.msdn.microsoft.com/Forums/sharepoint/en-US/a0d08e98-1fd6-42cf-b738-6ba3df082210/search-account-got-insufficient-sql-database-permissions-for-user-execute-permission-was-denied?forum=sharepointadmin
    http://sharepointpaul.blogspot.jp/2013/09/resolution-of-sharepoint-event-id-5214.html
    Thanks
    Daniel Yang
    TechNet Community Support

  • Error when save date in SQL database

    Hi all,
    I try to save a date in to a sql database, but when I run my program I get following error:
    Invalid object 11/21/05 9:32 AM of type java.sql.Date assigned to host variable 3.
    In my database table I have a fields date (type Data)en time type(Time).
    this is the source:
         public void createTitle(int title_id, String description, String user)
              throws SQLException {
              // TODO Auto-generated method stub
              Date date = new Date();
              String dateFormat = "YYYY-MM-DD";
              String timeFormat = "HH:mm:ss";
              SimpleDateFormat ddf = new SimpleDateFormat(dateFormat);
              SimpleDateFormat tdf = new SimpleDateFormat(timeFormat);
              String det = ddf.format(date);
              String creationTime = tdf.format(date);
              PreparedStatement newTitle = conn.prepareStatement("INSERT INTO BE_TITLE (TITLE_ID, DESCRIPTION, DATE_CREATE, TIME_CREATE, USER_CREATE) VALUES (?,?,?,?,?)");     
              try{
                   newTitle.setInt(1,title_id);
                   newTitle.setString(2,description);
                   newTitle.setDate(3, new java.sql.Date(date.getTime()));
                   newTitle.setTime(4, new java.sql.Time(date.getTime()));
                   newTitle.setString(5, user);
                   newTitle.executeUpdate();
              }finally{
                   newTitle.close();
    Can someone give me a solution for this error
    Richard

    Hi Richard,
    in your application, you atempt to insert a java.sql.Date value into a DATE column, which has non-zero time-components. Open SQL rejects such an atempt.
    Please allow me to quote from the OSS message (452727 2004), which Detlev refers to in his quote:
    <i>
    about one year ago, we had a very lengthy discussion on this issue - not only internally at SAP but also with Jonathan Bruce, the specification lead for JBDC at Sun.                                                                               
    We have the following dillema:                                                                               
    The javadoc of the constructor java.sql.Date(long) states:                                                                               
    "If the given milliseconds value contains time information, the driver will set the time components to the time in the default time zone (the time zone of the Java virtual machine running the application) that corresponds to zero GMT."                                                                               
    With Jonathan Bruce, we tried to clarify when this normalization should take place.                                                               
    1) Source code inspection shows that it is not the class java.sql.Date the performs the normalization.                                                                               
    2) The Javadoc states that the JDBC driver should perform the             
    normalization. If it is the driver, more open questions arrise:                                                                               
    a) When shall the normalization be performed?                             
       - if a setDate method is invoked?                                      
       - only for INSERT and update statements or also for all other parameters?                                                               
       - shall the getDate method perform the normalization?                  
       - what about set/getObject?                                            
       - the java Docs of set/getDate don't tell anything about normalization.                                                            
       - according to which time zone shall the value be normalized?          
         - UTC                                                                
         - vm time zone?                                                      
         - database time zone?                                                                               
    3) If an automatic normalization is performed, data will be modified upon insertion (or extraction) from the database. This means that we cannot quarantee if we insert and Date object retrieve it again that the two objects will be equal.
    In the discussion with Jonathan, we couldn't clarify the issue. However, Jonathan promised that the JDK should be modified in such a way that the JDK is responsible for the normalization. (So far such a change has not been done).                                                                               
    In order to be on the safe side, SAP chose that Open SQL should reject any Date values that are not normalized. This is what we have          
    implemented and documented. I agree, that in this repect, Open SQL is stricter than the JDBC standard demands.                                                                               
    To help the application with the task to normalize a Date (and Time) object, we offer the helper class                                      
    com.sap.sql.DateTimeNormalizer which you find in opensql.jar.                                                                               
    To come to a conclusion, the bevavoir you observed is from our side not a bug of Open SQL but a feature. Please use DateTimeNormalizer to normalize Date objects appropriately.                                  
    </i>
    I hope, you can follow our reasoning.
    Best regards,
    Adrian

  • Execution takes time for the report in Application

    I have developed a report for payroll runresult.Its execution time is 3 minutes,but when i run in application it takes half an hour
    The query i use is...
    SELECT round((sum(nvl(DECODE(PET.element_name,'Basic Salary',ROUND(prrv.result_value,3),0),0))+
    sum(nvl(DECODE(PET.element_name,'House Rent Allowance',ROUND(prrv.result_value,3),0),0))+
    sum(nvl(DECODE(PET.element_name,'Transport Allowance',ROUND(prrv.result_value,3),0),0))+
    sum(nvl(DECODE(PET.element_name,'Petrol Allowance',ROUND(prrv.result_value,3),0),0))+
    sum(nvl(DECODE(PET.element_name,'Vehicle Allowance',ROUND(prrv.result_value,3),0),0)) +
    sum(nvl(DECODE(PET.element_name,'Additional HRA',ROUND(prrv.result_value,3),0),0)) +
    sum(nvl(DECODE(PET.element_name,'Other Allowance',ROUND(prrv.result_value,3),0),0))+
    sum(nvl(DECODE(PET.element_name,'Carried Balance',ROUND(prrv.result_value,3),0),0))+
    sum(nvl(DECODE(PET.element_name,'Mobile Allowance',ROUND(prrv.result_value,3),0),0)) +
    sum(nvl(DECODE(PET.element_name,'Fixed Over Time',ROUND(prrv.result_value,3),0),0)) +
    sum(nvl(DECODE(PET.element_name,'Rent Deduction',ROUND(prrv.result_value,3),0),0)) -
    sum(nvl(DECODE(PET.element_name,'Vehicle Allowance Deducted',ROUND(prrv.result_value,3),0),0)) -
    sum(nvl(DECODE(PET.element_name,'HRA Deducted',ROUND(prrv.result_value,3),0),0)) -
    sum(nvl(DECODE(PET.element_name,'Rent Deduction',ROUND(prrv.result_value,3),0),0)) -
    sum(nvl(DECODE(PET.element_name,'Insurance Deduction',ROUND(prrv.result_value,3),0),0)) +
    sum(nvl(DECODE(PET.element_name,'Medical Insurance',ROUND(prrv.result_value,3),0),0)) +
    sum(nvl(DECODE(PET.element_name,'Family Passage Allowance',ROUND(prrv.result_value,3),0),0)) +
    sum(nvl(DECODE(PET.element_name,'Salik Allowance',ROUND(prrv.result_value,3),0),0)) -
    sum(nvl(DECODE(PET.element_name,'Family Insurance Deduction',ROUND(prrv.result_value,3),0),0)) +
    sum(nvl(DECODE(PET.element_name,'Fixed Incentive',ROUND(prrv.result_value,3),0),0)) +
    sum(nvl(DECODE(PET.element_name,'Company Accomodation Provided',ROUND(prrv.result_value,3),0),0))-
    sum(nvl(DECODE(PET.element_name,'Mess Deduction',ROUND(prrv.result_value,3),0),0))+
    sum(nvl(DECODE(PET.element_name,'Normal Overtime',ROUND(prrv.result_value,3),0),0))+
    sum(nvl(DECODE(PET.element_name,'Special Overtime',ROUND(prrv.result_value,3),0),0))+
    sum(nvl(DECODE(PET.element_name,'Telephone Allowance',ROUND(prrv.result_value,3),0),0))+
    sum(nvl(DECODE(PET.element_name,'Other Allowance',ROUND(prrv.result_value,3),0),0))+
    sum(nvl(DECODE(PET.element_name,'Other Earnings',ROUND(prrv.result_value,3),0),0))-
    sum(nvl(DECODE(PET.element_name,'Other Deductions',ROUND(prrv.result_value,3),0),0)) -
    sum(nvl(DECODE(PET.element_name,'Salary Advance Recovery',ROUND(prrv.result_value,3),0),0))-
    sum(nvl(DECODE(PET.element_name,'Air Ticket Refund',ROUND(prrv.result_value,3),0),0))-
    sum(nvl(DECODE(PET.element_name,'Vehicle Fines and Charges',ROUND(prrv.result_value,3),0),0))-
    sum(nvl(DECODE(PET.element_name,'Abroad Emergency Leave',ROUND(prrv.result_value,3),0),0))-
    sum(nvl(DECODE(PET.element_name,'Abscond Leave',ROUND(prrv.result_value,3),0),0))-
    sum(nvl(DECODE(PET.element_name,'Air Fare Paid',ROUND(prrv.result_value,3),0),0))-
    sum(nvl(DECODE(PET.element_name,'Recovery Carried Balance',ROUND(prrv.result_value,3),0),0))-
    sum(nvl(DECODE(PET.element_name,'Hajj Leave',ROUND(prrv.result_value,3),0),0))-
    sum(nvl(DECODE(PET.element_name,'Local Leave',ROUND(prrv.result_value,3),0),0))-
    sum(nvl(DECODE(PET.element_name,'Sick Leave Deduction',ROUND(prrv.result_value,3),0),0))-
    sum(nvl(DECODE(PET.element_name,'Sick Leave Without certificate',ROUND(prrv.result_value,3),0),0))-
    sum(nvl(DECODE(PET.element_name,'Round Off Deduction',ROUND(prrv.result_value,3),0),0))+
    sum(nvl(DECODE(PET.element_name,'Round Off Earnings',ROUND(prrv.result_value,3),0),0))),0)Net_Salary,papf.national_identifier groupcode,papf.FULL_NAME,paav.effective_date,
    haout.name costcentre,pg.name grade, pj.name designation ,paaf.ASSIGNMENT_NUMBER employee_number,decode(pop.org_payment_method_name,'RI_CASH','C','B')Payment
    FROM pay_run_results_v prrv,pay_run_results prr,pay_assignment_actions paa,per_all_assignments_f paaf,per_all_people_f papf,hr_all_organization_units_tl haout,
    per_grades pg, per_jobs pj,pay_assignment_actions_v paav,pay_personal_payment_methods_f ppp,pay_all_payrolls_f pap,pay_payroll_actions ppa,
    pay_org_payment_methods_f_tl pop,pay_element_types_f PET
    WHERE --prrv.assignment_action_id = 49449 and
    ppp.assignment_id(+) = paa.assignment_id
    AND pop.org_payment_method_id(+) = ppp.org_payment_method_id AND
    paa.ASSIGNMENT_ACTION_ID=prrv.ASSIGNMENT_ACTION_ID and paaf.ASSIGNMENT_ID=paa.ASSIGNMENT_ID
    and haout.organization_id=paaf.ORGANIZATION_ID
    and papf.EMPLOYEE_NUMBER=paaf.ASSIGNMENT_NUMBER
    AND prrv.run_result_id = prr.run_result_id
    AND ppa.effective_date BETWEEN pet.effective_start_date
    AND pet.effective_end_date
    AND prr.element_type_id = pet.element_type_id
    and prrv.CLASSIFICATION_NAME not like 'Info%'
    -- and prrv.element_name not like 'Gross Salary%'
    -- and prrv.element_name not like 'Annual%'
    and pj.job_id(+)=paaf.job_id
    and haout.name=nvl(:P_Dept,haout.name)
    and ppa.PAYROLL_ACTION_ID=paa.PAYROLL_ACTION_ID
    and pap.PAYROLL_ID=ppa.PAYROLL_ID
    and paav.ASSIGNMENT_ID=paaf.ASSIGNMENT_ID
    and paav.ASSIGNMENT_ACTION_ID=paa.ASSIGNMENT_ACTION_ID
    and pap.PAYROLL_NAME=NVL(:P_PAYROLL_NAME,pap.PAYROLL_NAME)
    --and papf.national_identifier = 03917
    and paav.effective_date >= TO_DATE ('01-' || :p_start_date, 'DD-MON-YYYY')
    AND paav.effective_date <= LAST_DAY (TO_DATE ('01-' || :p_start_date, 'DD-MON-YYYY'))
    -- and paav.DATE_EARNED=to_char(:p_date,'DD-MON-YYYY')
    and pg.grade_id(+)=paaf.grade_id group by papf.national_identifier,papf.FULL_NAME,paav.effective_date,
    haout.name ,pg.name , pj.name,paaf.ASSIGNMENT_NUMBER,pop.org_payment_method_name,papf.employee_number
    Cant analyse where it takes time...
    Regards,
    Mera

    I don't think that the "SUM" structure in the SELECT makes a difference in performance (And DO NOT use PL/SQL for this!) And as long as you SUM most PET.element_name values you do not need to filter by them.
    But, of course, the solution of "Sven W." would make the code MUCH nicer (took me a while to get it) and the "format" of the join conditions would take anyone ages to maintain that code!
    But your problem is that the execution differs massively right?
    Do you do the same thing? E.g. bind variables in the interactive part? Do the variables have the same value?
    Do you filter on the other variables like "P_DEPT" etc. ?
    "and haout.name=nvl(:P_Dept,haout.name)"
    The good thing is that you have an execution which makes you happy so you need to compare the SQL and the execution for both. Somewhere must be a difference. (Is the result the same by the way (just in case your parameters differ)?)
    If the results and parameters are the same, then do you not work with parameters in the "interactive" test and ORACLE is choosing a different execution plan as it knows more about your question. If that's the case, try to hint the query so that it's using the desired execution plan when you work with parameters.
    And how did you get that "execution plan" it's missing so much data and it's not formated. (put around it and check in "Preview" if it looks as desired)
    -- andy                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

  • SQL database login for 'SharePoint_Config_2010' on instance 'DB server' failed. Additional error information from SQL Server is included below.

    we formatted Database server and restored all the databases from the backup. but the sharepoint site is giving below error when i access the home page. 
    we have 3 application servers in that two are working fine,  but only one server getting the below error. 
    SQL database login for 'SharePoint_Config_2010' on instance 'db server' failed. Additional error information from SQL Server is included below.
    Login failed for user 'domain\user'.
    could you help what is the issue?

    Hi,
    According to your post, my understanding is that you failed to access sharepoint site with one application server.
    It was due to the MOM/SCOM Agent that has a SharePoint management pack installed, the agents Windows Service runs as ‘Local System’ and thus causes this.
    The workaround is to either kill the SCOM agent service, or set it to run as a dedicated service account.
    In addtion, I recommend to add the logon account of the SharePoint Timer service as a user of the SharePoint_Config database. Give role membership to the account db_owner.
    You also need to add the SharePoint accounts back into sql server.
    There are some similar articles for you to take a look at:
    http://blog.blksthl.com/2012/04/26/login-for-sharepoint_config-login-failed-for-user-domaincomputername-scom-agent/
    http://msdn.microsoft.com/en-us/library/jj551781(v=nav.70).aspx
    http://sharepoint.stackexchange.com/questions/51623/moved-sql-databases-for-sharepoint-2010-and-now-have-a-3351-error
    Best Regards,
    Linda Li
    Linda Li
    TechNet Community Support

  • How to connect sql database with Flash

    Dear Friends,
    Iam working on a quessinarie (assesment) for elearning software. i want to connect sql database with flash and save my records in back end like add, delete, modification of student result and details.
    Kindly help me to connect the sql database. or dot net server. I know how to connect PHP and mysql. But my client dont want php. only sql database and do the add, del, modificaiton.
    Thanks in advance,
    Syed Abdul Rahim

    unfortunately no, i do not know ASP well enough to give any code advice, other than the basic concept:  make a request on a server-side script (such as ASP) via URLLoader - the script receives variable values (if required), runs the DB queries you wish to invoke, and 'returns' the results - pick them up in the URLVariables class via the event handler in Flash and assign them to local properties if needed, etc.
    there are quite a few threads that you could find more specific information on the ASP side, here and elsewhere around the net.

Maybe you are looking for