Database time comparisons from RAT

I am using Real Application Testing to test moving to a new server and I am a little confused on the “database time” comparisons. I ran the capture for 5 minutes and 54 seconds, the database time for the CAPTURE shows 33 minutes and 29 seconds, when I replayed it the REPLAY shows database time for 6:09 but the actual duration was only 5:57.
So this tells me that it took about the same amount of time to play back the workload but I don’t understand the database time. Can somebody help me out with this? Thanks.

I have not done RAT but my guess is the CAPTURE time is the cumulative time consumed by sessions on database for the given workload.

Similar Messages

  • Currency Translation not working with "Time Reference from Variable"

    We have created several Currency Translations using RSCUR with a fixed "Key Date".  All have been working fine until we got the request to allow a variable date for the exchange rate date.  I have built a variable on 0DATE which I am putting a default date value into using user exit.  This default date is the SAME date as the fixed "Key Date" which we have been using.  Unfortunately when we use the "Time Reference from Variable" option and use this new variable, no currency conversion is taking place, instead it only shows the original currency values.  I am 100% sure that my variable is getting a value and it is the same date we are using for the fixed "Key Date" value so both methods should be using the exact same exchange rates.  The documentation that I can find states:
    ●      Time reference: The time reference for the currency translation can be either fixed or variable.
    If the time reference is fixed, the time at which the exchange rate is determined is independent of the data. You have the following options:
    ○       You can establish that the time reference be determined upon translation.
    ○       You can select the current date.
    ○       You can specify a fixed date as the key date.
    ○       You can specify any variable that exists for InfoObject 0DATE.
    ○       You can establish that the query key date be used. This is determined in the query settings.
    In my case I am only switching from option "You can specify a fixed date as the key date"=(Key Date) to "You can specify any variable that exists for InfoObject 0DATE"=(Time Reference from Variable).
    What am I missing?  Any help is appreciated.

    I neglected to mention that we are calling this currency translation through a WAD button using the SET_CURRENCY_TRANSLATION command.  I did do a test by applying the currency conversion within query key figure.  This seems to work.  So it appears the issue is with how the WAD is processing the variable, not the query.  When I display the variable in a drop down item within the WAD it does have the default value that I have assigned in user exit.

  • Approach to converting database character set from Western European to Unicode

    Hi All,
    EBS:12.2.4 upgraded
    O/S: Red Hat Linux
    I am looking for the below information. If anyone could help provide would be great!
    INFORMATION NEEDED: Approach to converting database character set from Western European to Unicode for source systems with large data exceptions
    DETAIL: We are looking to convert Oracle EBS database character set from Western European to Unicode to support Kanji characters. Our scan results show
    both “lossy (110K approx.)” and “truncation (26K approx.)” exceptions in the database which needs to be fixed before the database is converted to Unicode.
    Oracle Support has suggested to fix all open and closed transactions in the source Production instance using forms and scripts.
    We’re looking for information/creative approaches  who have performed similar exercises without having to manipulate data in the source instance.
    Any help in this regard would be greatly appreciated!
    Thanks for yourn time!
    Regards,

    There are two aspects here:
    1. Why do you have such large number of lossy characters? Is this data coming from some very old eBS release, i.e. from before the times of the Java applet interface to Oracle Forms?  Have you analyzed the nature of this lossy data?
    2. There is no easy way around truncation issues as you cannot modify eBS metadata (make columns wider). You must shorten or remove the data manually through the documented eBS interfaces. eBS does not support direct manipulation of data in the database due to complex consistency rules enforced by the application itself (e.g. forms).
    Thanks,
    Sergiusz

  • JAVA execution consumed significant database time

    Dear all,
    DB 11. on solaris..
    I have a time consuming package.. this package unloads data in the current db and loads data from another remote db in the same network. below is the recommendation from Oracle ADDM report.
    This session is a work flow session. What can I do for JAVA execution consumed significant database time.?
    JAVA execution consumed significant database time.
       Recommendation 1: SQL Tuning
       Estimated benefit is .5 active sessions, 43.98% of total activity.
       Action
          Tune the PL/SQL block with SQL_ID "6bd4fvsx8n42v". Refer to the "Tuning
          PL/SQL Applications" chapter of Oracle's "PL/SQL User's Guide and
          Reference".
          Related Object
             SQL statement with SQL_ID 6bd4fvsx8n42v.
             DECLARE job BINARY_INTEGER := :job; next_date DATE := :mydate;
             broken BOOLEAN := FALSE; BEGIN OWF_USER.START_METS_REFRESH(SYSDATE );
             :mydate := next_date; IF broken THEN :b := 1; ELSE :b := 0; END IF;
             END;Please advise
    Kai

    Hi Forstmann,
    Thanks for your update.
    Even i have collected ADDM report, extract of Node1 report as below
    FINDING 1: 40% impact (22193 seconds)
    Cluster multi-block requests were consuming significant database time.
    RECOMMENDATION 1: SQL Tuning, 6% benefit (3313 seconds)
    ACTION: Run SQL Tuning Advisor on the SQL statement with SQL_ID
    "59qd3x0jg40h1". Look for an alternative plan that does not use
    object scans.
    SYMPTOMS THAT LED TO THE FINDING:
    SYMPTOM: Inter-instance messaging was consuming significant database
    time on this instance. (55% impact [30269 seconds])
    SYMPTOM: Wait class "Cluster" was consuming significant database
    time. (55% impact [30271 seconds])
    FINDING 3: 13% impact (7008 seconds)
    Read and write contention on database blocks was consuming significant
    database time.
    NO RECOMMENDATIONS AVAILABLE
    SYMPTOMS THAT LED TO THE FINDING:
    SYMPTOM: Inter-instance messaging was consuming significant database
    time on this instance. (55% impact [30269 seconds])
    SYMPTOM: Wait class "Cluster" was consuming significant database
    time. (55% impact [30271 seconds])
    Any help from your side , please?
    Thanks,
    Sunand

  • Cluster multi-block requests were consuming significant database time

    Hi,
    DB : 10.2.0.4 RAC ASM
    OS : AIX 5.2 64-bit
    We are facing too much performance issues and CPU idle time becoming 20%.Based on the AWR report , the top 5 events are showing that problem is in cluster side.I placed 1st node AWR report here for your suggestions.
    WORKLOAD REPOSITORY report for
    DB Name DB Id Instance Inst Num Release RAC Host
    PROD 1251728398 PROD1 1 10.2.0.4.0 YES msprod1
    Snap Id Snap Time Sessions Curs/Sess
    Begin Snap: 26177 26-Jul-11 14:29:02 142 37.7
    End Snap: 26178 26-Jul-11 15:29:11 159 49.1
    Elapsed: 60.15 (mins)
    DB Time: 915.85 (mins)
    Cache Sizes
    ~~~~~~~~~~~ Begin End
    Buffer Cache: 23,504M 23,504M Std Block Size: 8K
    Shared Pool Size: 27,584M 27,584M Log Buffer: 14,248K
    Load Profile
    ~~~~~~~~~~~~ Per Second Per Transaction
    Redo size: 28,126.82 2,675.18
    Logical reads: 526,807.26 50,105.44
    Block changes: 3,080.07 292.95
    Physical reads: 962.90 91.58
    Physical writes: 157.66 15.00
    User calls: 1,392.75 132.47
    Parses: 246.05 23.40
    Hard parses: 11.03 1.05
    Sorts: 42.07 4.00
    Logons: 0.68 0.07
    Executes: 930.74 88.52
    Transactions: 10.51
    % Blocks changed per Read: 0.58 Recursive Call %: 32.31
    Rollback per transaction %: 9.68 Rows per Sort: 4276.06
    Instance Efficiency Percentages (Target 100%)
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    Buffer Nowait %: 99.87 Redo NoWait %: 100.00
    Buffer Hit %: 99.84 In-memory Sort %: 99.99
    Library Hit %: 98.25 Soft Parse %: 95.52
    Execute to Parse %: 73.56 Latch Hit %: 99.51
    Parse CPU to Parse Elapsd %: 9.22 % Non-Parse CPU: 99.94
    Shared Pool Statistics Begin End
    Memory Usage %: 68.11 71.55
    % SQL with executions>1: 94.54 92.31
    % Memory for SQL w/exec>1: 98.79 98.74
    Top 5 Timed Events Avg %Total
    ~~~~~~~~~~~~~~~~~~ wait Call
    Event Waits Time (s) (ms) Time Wait Class
    CPU time 18,798 34.2
    gc cr multi block request 46,184,663 18,075 0 32.9 Cluster
    gc buffer busy 2,468,308 6,897 3 12.6 Cluster
    gc current block 2-way 1,826,433 4,422 2 8.0 Cluster
    db file sequential read 142,632 366 3 0.7 User I/O
    RAC Statistics DB/Inst: PROD/PROD1 Snaps: 26177-26178
    Begin End
    Number of Instances: 2 2
    Global Cache Load Profile
    ~~~~~~~~~~~~~~~~~~~~~~~~~ Per Second Per Transaction
    Global Cache blocks received: 14,112.50 1,342.26
    Global Cache blocks served: 619.72 58.94
    GCS/GES messages received: 2,099.38 199.68
    GCS/GES messages sent: 23,341.11 2,220.01
    DBWR Fusion writes: 3.43 0.33
    Estd Interconnect traffic (KB) 122,826.57
    Global Cache Efficiency Percentages (Target local+remote 100%)
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    Buffer access - local cache %: 97.16
    Buffer access - remote cache %: 2.68
    Buffer access - disk %: 0.16
    Global Cache and Enqueue Services - Workload Characteristics
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    Avg global enqueue get time (ms): 0.6
    Avg global cache cr block receive time (ms): 2.8
    Avg global cache current block receive time (ms): 3.0
    Avg global cache cr block build time (ms): 0.0
    Avg global cache cr block send time (ms): 0.0
    Global cache log flushes for cr blocks served %: 11.3
    Avg global cache cr block flush time (ms): 1.7
    Avg global cache current block pin time (ms): 0.0
    Avg global cache current block send time (ms): 0.0
    Global cache log flushes for current blocks served %: 0.0
    Avg global cache current block flush time (ms): 4.1
    Global Cache and Enqueue Services - Messaging Statistics
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    Avg message sent queue time (ms): 0.1
    Avg message sent queue time on ksxp (ms): 2.4
    Avg message received queue time (ms): 0.0
    Avg GCS message process time (ms): 0.0
    Avg GES message process time (ms): 0.0
    % of direct sent messages: 6.27
    % of indirect sent messages: 93.48
    % of flow controlled messages: 0.25
    Time Model Statistics DB/Inst: PROD/PROD1 Snaps: 26177-26178
    -> Total time in database user-calls (DB Time): 54951s
    -> Statistics including the word "background" measure background process
    time, and so do not contribute to the DB time statistic
    -> Ordered by % or DB time desc, Statistic name
    Statistic Name Time (s) % of DB Time
    sql execute elapsed time 54,618.2 99.4
    DB CPU 18,798.1 34.2
    parse time elapsed 494.3 .9
    hard parse elapsed time 397.4 .7
    PL/SQL execution elapsed time 38.6 .1
    hard parse (sharing criteria) elapsed time 27.3 .0
    sequence load elapsed time 5.0 .0
    failed parse elapsed time 3.3 .0
    PL/SQL compilation elapsed time 2.1 .0
    inbound PL/SQL rpc elapsed time 1.2 .0
    repeated bind elapsed time 0.8 .0
    connection management call elapsed time 0.6 .0
    hard parse (bind mismatch) elapsed time 0.3 .0
    DB time 54,951.0 N/A
    background elapsed time 1,027.9 N/A
    background cpu time 518.1 N/A
    Wait Class DB/Inst: PROD/PROD1 Snaps: 26177-26178
    -> s - second
    -> cs - centisecond - 100th of a second
    -> ms - millisecond - 1000th of a second
    -> us - microsecond - 1000000th of a second
    -> ordered by wait time desc, waits desc
    Avg
    %Time Total Wait wait Waits
    Wait Class Waits -outs Time (s) (ms) /txn
    Cluster 50,666,311 .0 30,236 1 1,335.4
    User I/O 419,542 .0 811 2 11.1
    Network 4,824,383 .0 242 0 127.2
    Other 797,753 88.5 208 0 21.0
    Concurrency 212,350 .1 121 1 5.6
    Commit 16,215 .0 53 3 0.4
    System I/O 60,831 .0 29 0 1.6
    Application 6,069 .0 6 1 0.2
    Configuration 763 97.0 0 0 0.0
    Second node top 5 events are as below,
    Top 5 Timed Events
              Avg %Total
    ~~~~~~~~~~~~~~~~~~ wait Call
    Event Waits Time (s) (ms) Time Wait Class
    CPU time 25,959 42.2
    db file sequential read 2,288,168 5,587 2 9.1 User I/O
    gc current block 2-way 822,985 2,232 3 3.6 Cluster
    read by other session 345,338 1,166 3 1.9 User I/O
    gc cr multi block request 991,270 831 1 1.4 Cluster
    My RAM is 95GB each node and SGA is 51 GB and PGA is 14 GB.
    Any inputs from your side are greatly helpful to me ,please.
    Thanks,
    Sunand

    Hi Forstmann,
    Thanks for your update.
    Even i have collected ADDM report, extract of Node1 report as below
    FINDING 1: 40% impact (22193 seconds)
    Cluster multi-block requests were consuming significant database time.
    RECOMMENDATION 1: SQL Tuning, 6% benefit (3313 seconds)
    ACTION: Run SQL Tuning Advisor on the SQL statement with SQL_ID
    "59qd3x0jg40h1". Look for an alternative plan that does not use
    object scans.
    SYMPTOMS THAT LED TO THE FINDING:
    SYMPTOM: Inter-instance messaging was consuming significant database
    time on this instance. (55% impact [30269 seconds])
    SYMPTOM: Wait class "Cluster" was consuming significant database
    time. (55% impact [30271 seconds])
    FINDING 3: 13% impact (7008 seconds)
    Read and write contention on database blocks was consuming significant
    database time.
    NO RECOMMENDATIONS AVAILABLE
    SYMPTOMS THAT LED TO THE FINDING:
    SYMPTOM: Inter-instance messaging was consuming significant database
    time on this instance. (55% impact [30269 seconds])
    SYMPTOM: Wait class "Cluster" was consuming significant database
    time. (55% impact [30271 seconds])
    Any help from your side , please?
    Thanks,
    Sunand

  • Metrics "Database Time Spent Waiting (%)" is at ... for event class "Commit

    Hi guys
    I'm getting this warning every 20 minutes from Enterprise Manager 10g: Metrics "Database Time Spent Waiting (%)" is at 67.62891 for event class "Commit". My database is an Oracle 10g 10.2.0.4 (Linux x84_64) with dataguard, one physical and one standby networked at 100Mbit. I have 3 groups of 50M redo log with a low volume of transaction for now. Here are some metrics numbers:
    SQL> select sum(seconds_in_wait),event from v$session_wait group by event;
    SUM(SECONDS_IN_WAIT) EVENT
    121210 SQL*Net message from client
    0 Streams AQ: waiting for messages in the queue
    18 wait for unread message on broadcast channel
    6 LNS ASYNC end of log
    30 jobq slave wait
    37571 rdbms ipc message
    384 smon timer
    35090 pmon timer
    I tune the my listerners for Dataguard using the documentation. I don't know what to tune anymore, is someone has a clue.
    Thanks,
    Chris

    cprieur wrote:
    Metrics "Database Time Spent Waiting (%)" is at 67.62891 for event class "Commit"
    I'am also getting this message at a regular time:
    Metrics "Database Time Spent Waiting (%)" is at 64.09801 for event class "Network"
    The report, I was sent only to indicate the time spent on: SQL*Net message from client
    I believe that these metrics are explained here (near the bottom of the page):
    http://download.oracle.com/docs/cd/B19306_01/em.102/b25986/oracle_database.htm
    There are only two wait events in the Commit class (on 11.1.0.7):
    log file sync
    enq: BB - 2PC across RAC instances
    log file sync waits happen when sessions issue a COMMIT or ROLLBACK. Sessions will wait on this event until LGWR completes its writes to the redo logs (LGWR will likely wait on the event log file parallel wriite while waiting for the write to complete). I believe that your DataGuard configuration may contribute to the long waits experienced by LGWR, and it may be made worse by the network connection.
    http://download.oracle.com/docs/cd/B28359_01/server.111/b28294.pdf
    "Maximum availability - This protection mode provides the highest level of data protection that is possible without compromising the availability of a primary database. Transactions do not commit until all redo data needed to recover those transactions has been written to the online redo log and to at least one standby database."
    So, if Maximum availability is configured, the sessions will have to wait for the redo to be applied to the remote database.
    At the system-wide level you will almost always be able to ignore the SQL*Net message from client wait events. At the session level this wait event has a more significant meaning - the client computer is not actively waiting on a response from the database instance - the client is either sitting idle, or performing client-side processing.
    Sybrand, if he joins this thread, will likely be able to provide a more complete answer regarding DataGuard's contribution to the Commit wait class.
    Charles Hooper
    IT Manager/Oracle DBA
    K&M Machine-Fabricating, Inc.

  • System time is earlier than database time. (com.oracle.ovm.mgr.model.Model)

    I have recovered an OVM Manager to a different server using the "runInstaller --uuid", now I am getting the following error when I start the OVM Manager, which is version 3.0.3.
    ####<Apr 10, 2012 3:39:32 PM BST> <Info> <com.oracle.ovm.mgr.model.ModelEngine> <sbarber-vm5.uk.oracle.com> <AdminServer> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<anonymous>> <> <0000JQSgIqh9XbFpN0g8ye1FX4N^000002> <1334068772748> <BEA-000000> <Starting Model Engine Error
    java.lang.IllegalStateException: System time is earlier than database time. Please set system time after 2013-04-08 16:49:16 +0000
    at com.oracle.ovm.mgr.model.ModelEngine.checkSystemTime(ModelEngine.java:255)
    at com.oracle.ovm.mgr.model.ModelEngine.createModel(ModelEngine.java:201)
    at com.oracle.ovm.mgr.model.ModelEngine.start(ModelEngine.java:127)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at com.oracle.ovm.mgr.control.api.EngineManagerImpl.invokeMethod(EngineManagerImpl.java:335)
    at com.oracle.ovm.mgr.control.api.EngineManagerImpl.invokeMethod(EngineManagerImpl.java:307)
    at com.oracle.ovm.mgr.control.api.EngineManagerImpl.startEngine(EngineManagerImpl.java:512)
    at com.oracle.ovm.mgr.control.api.EngineManagerImpl.startModelEngine(EngineManagerImpl.java:463)
    at com.oracle.ovm.mgr.control.api.EngineManagerImpl.startEngines(EngineManagerImpl.java:547)
    at com.oracle.ovm.mgr.control.api.EngineManagerImpl.start(EngineManagerImpl.java:453)
    at com.oracle.ovm.mgr.control.service.VirtualizationService.startEngines(VirtualizationService.java:426)
    at com.oracle.ovm.mgr.control.service.VirtualizationService.start(VirtualizationService.java:125)
    at oracle.ovm.web.OVMManagerCoreServlet.init(OVMManagerCoreServlet.java:209)
    at weblogic.servlet.internal.StubSecurityHelper$ServletInitAction.run(StubSecurityHelper.java:283)
    at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
    at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:120)
    at weblogic.servlet.internal.StubSecurityHelper.createServlet(StubSecurityHelper.java:64)
    at weblogic.servlet.internal.StubLifecycleHelper.createOneInstance(StubLifecycleHelper.java:58)
    at weblogic.servlet.internal.StubLifecycleHelper.<init>(StubLifecycleHelper.java:48)
    at weblogic.servlet.internal.ServletStubImpl.prepareServlet(ServletStubImpl.java:539)
    at weblogic.servlet.internal.WebAppServletContext.preloadServlet(WebAppServletContext.java:1985)
    at weblogic.servlet.internal.WebAppServletContext.loadServletsOnStartup(WebAppServletContext.java:1959)
    at weblogic.servlet.internal.WebAppServletContext.preloadResources(WebAppServletContext.java:1878)
    at weblogic.servlet.internal.WebAppServletContext.start(WebAppServletContext.java:3153)
    at weblogic.servlet.internal.WebAppModule.startContexts(WebAppModule.java:1508)
    at weblogic.servlet.internal.WebAppModule.start(WebAppModule.java:482)
    at weblogic.application.internal.flow.ModuleStateDriver$3.next(ModuleStateDriver.java:425)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:52)
    at weblogic.application.internal.flow.ModuleStateDriver.start(ModuleStateDriver.java:119)
    at weblogic.application.internal.flow.ScopedModuleDriver.start(ScopedModuleDriver.java:200)
    at weblogic.application.internal.flow.ModuleListenerInvoker.start(ModuleListenerInvoker.java:247)
    at weblogic.application.internal.flow.ModuleStateDriver$3.next(ModuleStateDriver.java:425)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:52)
    at weblogic.application.internal.flow.ModuleStateDriver.start(ModuleStateDriver.java:119)
    at weblogic.application.internal.flow.StartModulesFlow.activate(StartModulesFlow.java:27)
    at weblogic.application.internal.BaseDeployment$2.next(BaseDeployment.java:636)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:52)
    at weblogic.application.internal.BaseDeployment.activate(BaseDeployment.java:205)
    at weblogic.application.internal.EarDeployment.activate(EarDeployment.java:58)
    at weblogic.application.internal.DeploymentStateChecker.activate(DeploymentStateChecker.java:161)
    at weblogic.deploy.internal.targetserver.AppContainerInvoker.activate(AppContainerInvoker.java:79)
    at weblogic.deploy.internal.targetserver.BasicDeployment.activate(BasicDeployment.java:184)
    at weblogic.deploy.internal.targetserver.BasicDeployment.activateFromServerLifecycle(BasicDeployment.java:361)
    at weblogic.management.deploy.internal.DeploymentAdapter$1.doActivate(DeploymentAdapter.java:51)
    at weblogic.management.deploy.internal.DeploymentAdapter.activate(DeploymentAdapter.java:200)
    at weblogic.management.deploy.internal.AppTransition$2.transitionApp(AppTransition.java:30)
    at weblogic.management.deploy.internal.ConfiguredDeployments.transitionApps(ConfiguredDeployments.java:240)
    at weblogic.management.deploy.internal.ConfiguredDeployments.activate(ConfiguredDeployments.java:169)
    at weblogic.management.deploy.internal.ConfiguredDeployments.deploy(ConfiguredDeployments.java:123)
    at weblogic.management.deploy.internal.DeploymentServerService.resume(DeploymentServerService.java:180)
    at weblogic.management.deploy.internal.DeploymentServerService.start(DeploymentServerService.java:96)
    at weblogic.t3.srvr.SubsystemRequest.run(SubsystemRequest.java:64)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:209)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:178)
    OVM Manager date and time are;
    $date
    Tue Apr 10 15:53:45 BST 2012
    The database is remote and it's system date and time is;
    $ date
    Wed Apr 10 07:36:34 PDT 2013
    Verified by;
    SQL> select to_char(sysdate,'DD-MON-YYYY HH24:MI:SS') from dual;
    TO_CHAR(SYSDATE,'DD-
    10-APR-2013 07:37:58
    Any help gratefully recieved.

    I wish you luck at running a weblogic application attaching to a database that is 8 time zones away. BST is GMT +1 and PDT is GMT-7 I hope you have plenty of bandwidth. I'd say at least the latency of the connection would be an issue.
    Regardless, time zones are based on hours apart. Your time difference is off by more than just "whole hours"? Its also not 8 hours difference. Its almost 11 hours. Also, the date output of you host says 2012. Surely that is mistake?

  • Time Comparison PHP

    Hi There
    I am trying to create a time comparison for a website I am building.
    I have the following so far
    <?php
    $date=$_POST ['daydropdown'];
    $month=$_POST ['monthdropdown'];
    $year=$_POST ['yeardropdown'];
    $datevalue="$date/$month/$year";
    $booking=strtotime ("now + 24 hours");
    if ($date_value > $booking)
    echo ("lovey");
    else
    echo ("tut tut");
    ?>
    which is pulling its date info from
              <form action="timetest.php" method="post">
              <table>
      <tr>
        <td class="tableheight">
          Party Size</td>
        <td>
          <select name="partysize" size="1" id="partysize">
          <option>1</option>
          <option selected="selected">2</option>
          <option>3</option>
          <option>4</option>
          <option>5</option>
          <option>6</option>
          <option>7</option>
          <option>8</option>
          <option>9</option>
          <option>10</option>
          <option>11</option>
          <option>12</option>
          <option>13</option>
          <option>14</option>
          <option>15</option>
          <option>16</option>
          <option>17</option>
          <option>18</option>
          <option>19</option>
          <option>20</option>
        </select>
        </td>
      </tr>
      <tr>
        <td class="tableheight">Date</td>
        <td>
    <select name="daydropdown" id="daydropdown">
    </select>
    <select name="monthdropdown" id="monthdropdown">
    </select>
    <select name="yeardropdown" id="yeardropdown">
    </select>
    </td>
      </tr>
      <tr>
        <td class="tableheight">Time</td>
        <td><select name="time">
          <option>12 noon</option>
          <option>12:15pm</option>
          <option>12:30pm</option>
          <option>12:45pm</option>
          <option>1:00pm</option>
          <option>1:15pm</option>
          <option>1:30pm</option>
          <option>1:45pm</option>
          <option>2:00pm</option>
          <option>2:15pm</option>
          <option>2:30pm</option>
          <option>2:45pm</option>
          <option>3:00pm</option>
          <option>3:15pm</option>
          <option>3:30pm</option>
          <option>3:45pm</option>
          <option>4:00pm</option>
          <option>4:15pm</option>
          <option>4:30pm</option>
          <option>4:45pm</option>
          <option>5:00pm</option>
          <option>5:15pm</option>
          <option>5:30pm</option>
          <option>5:45pm</option>
          <option>6:00pm</option>
          <option>6:15pm</option>
          <option>6:30pm</option>
          <option>6:45pm</option>
          <option>7:00pm</option>
          <option>7:15pm</option>
          <option>7:30pm</option>
          <option>7:45pm</option>
          <option>8:00pm</option>
          <option>8:15pm</option>
          <option>8:30pm</option>
          <option>8:45pm</option>
          <option>9:00pm</option>
          <option>9:15pm</option>     
    </select></td>
      </tr>
      <tr>
        <td class="tableheight">First Name</td>
        <td><input name="firstname" type="text" placeholder="Your First Name" /></td>
      </tr>
      <tr>
        <td class="tableheight">Surname</td>
        <td><input name="surname" type="text" placeholder="Your Last Name"/></td>
      </tr>
      <tr>
        <td class="tableheight">Email Address</td>
        <td><input name="email" type="email" placeholder="Your Email Address"/></td>
      </tr>
      <tr>
        <td class="tableheight">Confirm Email </td>
        <td><input name="confirm" type="email"  placeholder="Confirm Email"/></td>
      </tr>
      <tr>
        <td class="tableheight">Contact Number</td>
        <td><input name="contactnumber" type="text" placeholder="Your Contact Number"/></td>
      </tr>
      <tr>
        <td style="padding-top:0px"><input name="Book Now" type="submit" value="Book Now" /></td>
        <td></td>
      </tr>
    </table>
              </form>
              <script type="text/javascript">
    //populatedropdown(id_of_day_select, id_of_month_select, id_of_year_select)
    window.onload=function(){
    populatedropdown("daydropdown", "monthdropdown", "yeardropdown")
    </script>
    What I want to happen is if a booking for a table is made for more than 24hours in advance then its fine and sends the reservation to the restaurant aka "Lovely" where as if its for within the next 24 hours then "tut tut" they are told to ring the restaurant.
    Does anyone have any pointers for me please.
    G

    You need to change the $datevalue variable to a timestamp like you did with $booking.
    $dateTried = strtotime($datevalue); 
    Then you can compare them.
    if ($dateTried < $booking) {      // this is what to do if the time chosen was within 24 hours } else {      // this is what to do if the time chosen was a valid value (24 hours or more from current time) } 
    BTW, it looked like your form fields are populated using JavaScript. If I were you I would switch that to be PHP driven. The code will be very similar in PHP as your JavaScript code, but it will work for all your visitors, not just the ones that have JavaScript enabled.
    -Jason
    Sorry the code didn't line up. I used the wrong code insert options.
    Message was edited by: UteFanJason

  • Wait class 'commit' consuming significant database time.

    Hi
    My awr report was showing that the log file sync as the top wait event.I can also see additional message saying that wait class 'commit' consuming significant database time.Can any one suggest me what are the tuning things i need to consider for this wait class 'LOG FILE SYNC'
    Thanks

    Be very careful about this.
    Follow this only if you can afford to lose some data in case of instance failure
    (eg death of the instance from a bug, server panic/reboot, power failure etc).
    Oracle's normal behaviour is to guarantee that every committed transaction
    IS available by ensuring that it is in the redologs and reapplying it if necessary
    in case of an instance failure and recovery or media recovery.
    A Commit NOWAIT means that there is a possibility, however slight, that
    the last few transaction(s) might not have gotten into the redo logs at the time
    of instance failure.
    Your application / analysts must be able to identify transactions that are 'lost'
    and reapply them after you restart a crashed instance.

  • A new database is available from iCloud

    What does this message mean and should I update to the newest database?  HJZ

    I am experiencing a similiar problem. Nearly every time I open my Runmeter app on my iPhone I get the message "iCloud update. A new database is available from iCloud." I now have more than 20 different databases, each with my name followed by a serial number as Harry described.
    Details:
    App: Runmeter Elite version 10.3.6 (14824) by Abvio Inc.
    Device: iPhone 5 model ME486LL/A
    Operating System: iOS 8.1.2
    Carrier: T-Mobile 18.1
    Firmware: 7.03.00
    Anyone have any suggestions?

  • Time reference from variable in crcy transl. type

    Hi guys,
    does anybody know, where to create the variables displayed in RSCUR, when I want to create a crcy. transl. type with time reference from variable?
    Thanks in advance, Clemens

    Hi,
    With  step by step :
    http://www.sapdb.info/wp-content/uploads/2008/11/bw-howtousevariablestimereferencesincurrencyconversion.pdf
    with  steps and Definitions :
    http://www.scribd.com/doc/7061035/How-to-Use-Variables-for-Currency-Conversion
    The time reference for the currency translation can be either fixed or variable.
    If the time reference is fixed, the time at which the exchange rate is determined is independent of the data. You have the following options:
    ○       You can establish that the time reference be determined upon translation.
    ○       You can select the current date.
    ○       You can specify a fixed date as the key date.
    ○       You can specify any variable that exists for InfoObject 0DATE.
    ○       You can establish that the query key date be used. This is determined in the query settings.
    If the time reference is variable, the time at which the exchange rate is determined is decided by a time characteristic value.
    ○       A variable time reference can, for example, be determined using the end or start of a fiscal year or calendar year, a period and a month u2013 or even to the exact day. The following standard time characteristics are available: 0FISCYEAR, 0FISCPER, 0CALYEAR, 0CALQUARTER, 0CALMONTH, 0CALWEEK and 0CALDAY.
    ○       The time reference can also be determined using a customer-specific InfoObject of type date (for example, trading day). Note that this InfoObject has to have the same properties as the standard InfoObject selected in the variable time reference or must reference it.
    ○       You can specify an InfoSet: This is only necessary if you are determining the target currency using an InfoObject that exists in the InfoSet more than once. You enter the field alias so that the InfoObject can be specified uniquely.
    In the Business Explorer, you can only set time variable currency translations in query definition, and not for an executed query. See also Setting Variable Target Currency in the Query Designer.
    ○       In the Time Adjustment field of type INT4, you can specify whole numbers with a +/- sign.
    ○       In the Time Adjustment from Variable field, you can specify formula variables (1FORMULA). As these values of the variables may have to be whole numbers, they are rounded to whole numbers (integers) where necessary.
    The time adjustment (regardless of whether it is fixed or variable) is always related to the InfoObject specified under Variable Time Reference.
    Example of a time adjustment of -3:
    The variable time reference is To the exact day and the InfoObject under Variable Time Reference is 0CALDAY. Instead of 07.11.2006, after the time adjustment, 04.11.2006 is used for the translation.
    The variable time reference is End of Week and the InfoObject under Variable Time Reference is 0CALWEEK. Instead of 52.2004, the week 49.2004 is calculated and the end date is calculated from this.
    For more info pls refer this link:
    http://help.sap.com/saphelp_nw2004s/helpdata/EN/80/1a6806e07211d2acb80000e829fbfe/content.htm
    I hope these will helps you.
    Thanks .
    Hema

  • Create the 2nd Standby database on Linux from 1st standby database on wind

    Hi Experts
    I want to create the 2nd Standby database on Linux from 1st standby database on window.....
    my 2nd standby database is on mount state, when i try to recover by the given command then its giving errror.
    rman > DUPLICATE TARGET DATABASE FOR STANDBY DORECOVER NOFILENAMECHECK;
    My question is am I able to created it on cross platform....

    thanks a lot for your time and help Mahir!!!
    i tried through goldengate but i didn't able due to clob, long datatype....
    check the below link where i ask the question.....
    initial load failed to transfer the clob data

  • Contention on index block splits  consuming significant database time

    Hi Guys,
    can anybody suggest on how to remove Contention on index block splits,this is giving so many issues on my production DB,the CPU usage shots up and application hangs for few minutes.
    DB is 10.2.0.3 and OS is IBM AIX 5.3

    I found this.. it might be useful
    One possibility is that this is caused by shared CBC latching peculiarities:
    1) during normal selects your index root block can be examined under a
    shared cache buffers chains latch.
    So as long as everybody is only reading the index root block, everybody can
    do it concurrently (without pinning the block). The "current holder count"
    in the CBC latch structure is just increased by one for every read only
    latch get and decreased by one on every release. 0 value means that nobody
    has this latch taken currently.
    Nobody has to wait for others for reading index root block in all read only
    case. That greatly helps to combat hot index root issues.
    2) Now if a branch block split happens a level below the root block, the
    root block has to be pinned in exclusive mode for reflecting this change in
    it. In order to pin a block you need to get the corresponding CBC latch in
    exclusive mode.
    If there are already a bunch of readers on the latch, then the exclusive
    latch getter will just flip a bit in the CBC latch structure - stating it's
    interest for exclusive get.
    Every read only latch get will check for this bit, if it's set, then the
    getters will just spin instead, waiting this bit to be cleared (they may
    yield or sleep immediately as well, I haven't checked). Now the exclusive
    getter has to spin/wait until all the shared getters have released the latch
    and the "current holder count" drops to zero. Once it's zero (and the getter
    manager to get on to CPU) it can get the latch, do its work and release the
    latch.
    During all that time starting from when the "exclusive interest" bit was
    set, nobody could access this indexes root block except the processes which
    already had the latch in shared mode. Depending on latch spin/sleep strategy
    for this particular case and OSD implementation, this could mean that all
    those "4000 readers per second" start just spinning on that latch, causing
    heavy spike in CPU usage and they all queue up.
    How do diagnose that:
    You could sample v$latch_misses to see whether the number of "kcbgtcr:
    kslbegin shared" nowaitfails/sleeps counter takes an exceptional jump up
    once you observe this hiccup.
    How to fix that once diagnosed:
    The usual stuff, like partitioning if possible or creating a single table
    hash cluster instead.
    If you see that the problem comes from excessive spinning, think about
    reducing the spinning overhead (by reducing spincount for example). This
    could affect your other database functions though..
    If you can't do the above - then if you have off-peak time, then analyse
    indexes (using treedump for start) and if you see a block split coming in a
    branch below root block, then force the branch block to split during
    off-peak time by inserting carefully picked values into the index tree,
    which go exactly in the range which cause the proper block to split. Then
    you can just roll back your transaction - the block splits are not rolled
    back nor coalesced somehow, as this is done in a separate recursive
    transaction.
    And this
    With indexes, the story is more complicated since you can't just insert a
    row into any free block available like with tables. Multiple freelists with
    tables help us to spread up inserts to different datablocks, since every
    freelist has its distinct set of datablocks in it. With indexes, the
    inserted key has to go exactly to the block where the structure of b?tree
    index dictates, so multiple freelists can't help to spread contention here.
    When any of the index blocks has to split, a new block has to be allocated
    from the freelist (and possibly unlinked from previous location in index),
    causing an update to freelist entry in segment header block. Now if you had
    defined multiple freelists for your segment, they'd still remain in the
    single segment header block and if you'd have several simultaneous block
    splits, the segment header would become the bottleneck.
    You could relieve this by having multiple freelist groups (spreading up
    freelists into multiple blocks after segment header), but this approach has
    it's problems as well - like a server process which maps to freelist group 1
    doesn't see free blocks in freelist group 2, thus possibly wasting space in
    some cases...
    So, if you have huge contention on regular index blocks, then you should
    rethink the design (avoid right hand indexes for example), or physical
    design (partition the index), increasing freelists won't help here.
    But if you have contention on index segment's header block because of block
    splits/freelist operations, then either partition the index or have multiple
    freelist groups, adding freelists again won't help here. Note that adding
    freelist groups require segment rebuild.

  • Database Time Spent waiting issue

    Hi;
    I got this message in OEM:
    Metrics "Database Time Spent Waiting (%)" is at 62.00034 for event class "Other", under category Waits by Wait Class.
    When I hit over the message and then go to Additional Advice, under Impact(%) I got the message:
    ADDM did not find any problems during this period
    And under Informational Findings the message is:
    There was no significant database activity to run the ADDM.
    I am seeing this message in the last 3 days, however, apparently there's not problem. But the message still is there. My doubt is, is this a problem to watch? How can I delete this message?
    thanks for your advice
    Regards
    Al

    "SQL*Net message from client" event wait time is accumulated while your clients are idling (doing nothing but keeping the connection open.) If you have no apparent performance issues, this event can be safely ignored.
    Regards,
    Vladimir M. Zakharychev

  • Database Time Spent Waiting (%): Wait Class Other nearly 100% all the time

    Don't understand why OEM is reporting that Database Time Spent Waiting (%): Wait Class Other is nearly 100% all the time. Database 10.1.0.4 just installed on Linux(RHEL4AS) and nobody use it for now except OEM and me for admin purpose.
    Any clue for that problem ?
    Regards
    Nicolas

    Seems like you are not the first to see this kind of behaviour.
    I've found another similar thread on metalink. I can't say the answer is terribly helpful, but thought you might be interested anyway:
    From: Jose Ramón Tourón 14-Sep-07 08:34
    Subject: Database Time Spent Waiting (%) at 100 in event class Commit
    RDBMS Version: Oracle 10g r2
    Operating System and Version: Suse Enterprise Linux 10
    Error Number (if applicable):
    Product (i.e. SQL*Loader, Import, etc.): database core
    Product Version: 10gR2
    Database Time Spent Waiting (%) at 100 in event class Commit
    Hi everyone, yesterday we create a new database instance with dbca, the creation process was ok, and the two instances are running ok, database stops and starts without any problem, and listeners are ok. In the enterprise manager of this new instance we found this message:
    Database Time Spent Waiting (%) at 100 in event class Commit, this event happend every 1 or 2 minutes sometimes at 100, the next 40%, the next 98, ... and so on.
    Do you know what's happend in this instance?
    Thanks in advance to everyone
    Santiago Pérez
    From: Oracle, Helmut Pfau 14-Sep-07 12:52
    Subject: Re : Database Time Spent Waiting (%) at 100 in event class Commit
    From Oracle Database Reference Manual:
    Commit
    This wait class only comprises one wait event - wait for redo log write confirmation after a commit (that is, 'log file sync')
    So you can't write fast enough into your log files.
    Did you check the frequency of log switches?
    From: Metalink TCS User Group TCS Uruguay 14-Sep-07 15:52
    Subject: Re : Database Time Spent Waiting (%) at 100 in event class Commit
    Hi José! Please don't get anxious because of this: Wait time must be SOMEWHERE, there's a saying "An OLTP DB is only as fast as its redo logs", but if you are not having any performance problem you don't need to do anything special.
    You say you've just created the DB. Now make it DO something: put it to the test by simulating production conditions as closely as you can, and after some hours ask the users whether there is some problem. If there is, take a look at the wait statistics... you'll probably see many other top events before this one!
    Bruno abate_at_adinet.com.uy

Maybe you are looking for

  • Syncing apps between devices

    I'm having difficulty syncing some apps between devices so that data is saved. I recently purchased an iPad mini for my daughter and want her to have some of her existing game apps to play, and then delete from my iPad. Anyone encountered the same, o

  • MFC application with Sql Server Database

    Hello! I have made an application (using MFC ODBC costumer) and i want to make this program run in every computer. Is there any way to run it without installing sql server in every computer? Thank you!

  • Problem installing sound driver on Lenovo G550 - Windows XP

    I have installed windows XP on my lenovo G550. As shown on lenovo site, i tried to install"Conexant Audio Driver" for my audio driver but it is failing. I opened readme file it says somthing like "Change Point: Add .Net3.5 in driver package"- Does th

  • How to make an entire word spin/rotate?

    Is it possible to make an entire text word spin round. Moreover, can I zoom the word in whilst spinning?

  • Report with graph

    Hi all. I've 2 problems: I need to do a report display a graph with more than 6 period... I saw that the "GRAPH_MATRIX*" functions  use up to 6 elements for dimensions... I need at least 12 elements... The graph must have 2 lines and the quantities f