SQL History log

How do you sort the SQL History log entries in ASC or DSC timestamp order?
I've tried clicking on the "Timestamp" column and it changes, but it seems to be a problem with handling dates with a 1 digit day along with a 2 digit day( probably same will happen with 1 and 2 digit months ), for example 4/7, 4/6, 4/5 come before 4/14, 4/13, 4/12.
tia

The SQL History sorts on the text value of the timestamp column, rather than the date value of the timestamp column and has been doing this as long as I have been using it. This issue has been posted several times (not recently from memory), but I don't know how high up the priorities it has ended up.

Similar Messages

  • If I resize job history log to see more entries listed under job history, will cause sql server agent to restart or any downtime?

    Here is I am going to do:
    Using SQL Server Management Studio
    To resize the job history log based on raw size
    In Object Explorer, connect to an instance of the SQL Server Database Engine, and then expand that instance.
    Right-click SQL Server Agent, and then click Properties.
    Select the History page, and then confirm that Limit size of job history logis checked.
    In the Maximum job history log size box, enter the maximum number of rows the job history log should allow.
    In the Maximum job history rows per job box, enter the maximum number of job history rows to allow for a job.

    this change does not require SQL agent restart
    Please click the Mark as answer button and vote as helpful if this reply solves your problem

  • SQL Developer SQL History no longer working 3.2.20.09.87

    Hi guys, The SQL History (F8) option has started to no longer work within my copy of OSD.
    it was working at one point but has gone away.
    I have confirmed that "Enable Local History" is enabled and has been set for 30 days history and 50 max revisions
    I have confirmed that I still have a SqlHistory.xml file located in my user\operator\AppData\Roaming\SQL Developer\ directory with entries in it (6933 rows of data)
    This started to happen after I started to play with the Garbage Collection settings to reduce the memory footprint of OSD.
    I have since restored all of my backed up delivered settings, rebooted and the SQL History will not show up again.
    My snippets and custom reports are working and I can still see the log detail where I used to see log and sql history options at the bottom of OSD.
    I'm using Windows 7 and can provide an OSD dump if it helps.
    Thanks...
    Tom
    Edited by: ERPDude on Mar 6, 2013 10:03 AM

    Hi Tom,
    Be careful about which preference settings relate to the functionality in question.
    1. Tools > Preferences > Environment > Local History > Enable relates to a worksheet's History tab when you File > Open
    2. Tools > Preferences > Database > Worksheet > SQL History Limit relates to View > SQL History pane.
    You do not say how you backed up and restored your settings during all your experimentation, but let's assume whatever you did was valid. Whenever some UI component mysteriously disappears, one thing I always check (usually just delete) is the windowinglayout.xml in system3.2.20.09.87\o.ide.11.1.1.4.37.59.48 of my user settings.
    For example, I can use the SQL History pane's tab context menu to Float it, then adjust it's size by dragging it's edges down to a grey square a few pixels wide/high. Close SQL Developer, reopen, View > SQL History. Darn! Where did it go?
    Hope this helps,
    Gary

  • How to view executed sql history?

    I have encountered some strange errors in my productive database. I want to view the sql execution logs. but I don't know how to view it. is there any good methods to check the sql execution history?

    The core dump trace file will have information you need for troubleshooting. Most of time including the SQL that caused the problem if it was caused by certain SQL.
    Check your core dump destination. Core dump is not usual in Oracle operation, it's worth to submit a SR .

  • BUG in purging sql history entries...

    Hi folks,
    I am using 1.5 and have been using also the versions before. Great that the sql-history window is dockable now, much more useful this way...But...
    it is still totally ununderstanble to me, with wich rule entries in the sql-history are being purged. I just open the sqldev again, and see almost entries removed. When does that happen, is it possible to configure this setting?

    Hi guys,
    yes, I mean an arbitrary purge. Like, you finish your job for today, close sqldev, go home, come back the next day and only 4 or 5 entries of the log are still there... (???)
    No problem, I have full understanding for an algorithm for purging entries older than two or four days, or something similar. But until now, there was no repliccable rule in that process!
    By the way, I cannot find a sqlhistory.xml in my install-directory...

  • Extract SQL history from 10046 trace files

    Hi all,
    I need to extract the complete sql history from sql trace files to "debug" a client application.
    I know I can read the raw trc file and rebuild the sql history looking for the PARSING / EXEC / FETCH entries.
    However, this is a very long and boring manual task: do you know if there is some free tool to automate this task?
    thanks
    Andrea

    user585511 wrote:
    I agree that the 10046 trace captures everything. If I do read the raw trc file I see the DML. The problem is that tkprof's record does not record the DML (maybe it thinks that some DML is recursive sql and it gets misleaded... I am not sure) so I am looking for an alternate tool to process 10046 trace files
    Regards
    AndreaReally?
    Generate a trace of some dml:
    oracle:orcl$
    oracle:orcl$ sqlplus /nolog
    SQL*Plus: Release 11.2.0.1.0 Production on Thu May 16 08:28:55 2013
    Copyright (c) 1982, 2009, Oracle.  All rights reserved.
    SQL> conn snuffy/snuffy
    Connected.
    SQL> alter session set tracefile_identifier = "snuffy_session";
    Session altered.
    SQL> alter session set events '10046 trace name context forever, level 12';
    Session altered.
    SQL> insert into mytest values (sysdate);
    1 row created.
    SQL> commit;
    Commit complete.
    SQL> ALTER SESSION SET EVENTS '10046 trace name context off';
    Session altered.
    SQL> exitrun tkprof on the trace
    oracle:orcl$ ls -l $ORACLE_BASE/diag/rdbms/$ORACLE_SID/$ORACLE_SID/trace/*snuffy
    *.trc
    -rw-r----- 1 oracle asmadmin 3038 May 16 08:29 /u01/app/oracle/diag/rdbms/orcl/orcl/trace/orcl_ora_4086_snuffy_session.trc
    oracle:orcl$ tkprof /u01/app/oracle/diag/rdbms/orcl/orcl/trace/orcl_ora_4086_snu
    ffy_session.trc snuffy.rpt waits=YES sys=NO explain=system/halftrack
    TKPROF: Release 11.2.0.1.0 - Development on Thu May 16 08:31:32 2013
    Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.Look at the report:
    oracle:orcl$ cat snuffy.rpt
    TKPROF: Release 11.2.0.1.0 - Development on Thu May 16 08:31:32 2013
    Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.
    Trace file: /u01/app/oracle/diag/rdbms/orcl/orcl/trace/orcl_ora_4086_snuffy_session.trc
    Sort options: default
    count    = number of times OCI procedure was executed
    cpu      = cpu time in seconds executing
    elapsed  = elapsed time in seconds executing
    disk     = number of physical reads of buffers from disk
    query    = number of buffers gotten for consistent read
    current  = number of buffers gotten in current mode (usually for update)
    rows     = number of rows processed by the fetch or execute call
    SQL ID: 938dgt554gu98
    Plan Hash: 0
    insert into mytest           <<<<<<<<<<<<<<<< oh my!  Here is the insert statement
    values
    (sysdate)
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          1          5           1
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.00       0.00          0          1          5           1
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 86  (SNUFFY)
    Rows     Row Source Operation
          0  LOAD TABLE CONVENTIONAL  (cr=1 pr=0 pw=0 time=0 us)
    error during execute of EXPLAIN PLAN statement
    ORA-00942: table or view does not exist
    parse error offset: 83
    Elapsed times include waiting on following events:
      Event waited on                             Times   Max. Wait  Total Waited
      ----------------------------------------   Waited  ----------  ------------
      SQL*Net message to client                       1        0.00          0.00
      SQL*Net message from client                     1        3.35          3.35
    SQL ID: 23wm3kz7rps5y
    Plan Hash: 0
    commit
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          0          1           0
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.00       0.00          0          0          1           0
    Misses in library cache during parse: 0
    Parsing user id: 86  (SNUFFY)
    Elapsed times include waiting on following events:
      Event waited on                             Times   Max. Wait  Total Waited
      ----------------------------------------   Waited  ----------  ------------
      SQL*Net message to client                       2        0.00          0.00
      SQL*Net message from client                     2        4.72          8.50
      log file sync                                   1        0.00          0.00
    SQL ID: 0kjg1c2g4gdcr
    Plan Hash: 0
    ALTER SESSION SET EVENTS '10046 trace name context off'
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          0          0           0
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.00       0.00          0          0          0           0
    Misses in library cache during parse: 0
    Parsing user id: 86  (SNUFFY)
    OVERALL TOTALS FOR ALL NON-RECURSIVE STATEMENTS
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        3      0.00       0.00          0          0          0           0
    Execute      3      0.00       0.00          0          1          6           1
    Fetch        0      0.00       0.00          0          0          0           0
    total        6      0.00       0.00          0          1          6           1
    Misses in library cache during parse: 0
    Elapsed times include waiting on following events:
      Event waited on                             Times   Max. Wait  Total Waited
      ----------------------------------------   Waited  ----------  ------------
      SQL*Net message to client                       3        0.00          0.00
      SQL*Net message from client                     3        4.72         11.86
      log file sync                                   1        0.00          0.00
    OVERALL TOTALS FOR ALL RECURSIVE STATEMENTS
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        0      0.00       0.00          0          0          0           0
    Execute      0      0.00       0.00          0          0          0           0
    Fetch        0      0.00       0.00          0          0          0           0
    total        0      0.00       0.00          0          0          0           0
    Misses in library cache during parse: 0
        3  user  SQL statements in session.
        0  internal SQL statements in session.
        3  SQL statements in session.
        0  statements EXPLAINed in this session.
    Trace file: /u01/app/oracle/diag/rdbms/orcl/orcl/trace/orcl_ora_4086_snuffy_session.trc
    Trace file compatibility: 11.1.0.7
    Sort options: default
           1  session in tracefile.
           3  user  SQL statements in trace file.
           0  internal SQL statements in trace file.
           3  SQL statements in trace file.
           3  unique SQL statements in trace file.
          58  lines in trace file.
           8  elapsed seconds in trace file.
    oracle:orcl$

  • SQL History always emtpy or disabled

    When I started evaluating SQL developer, I used the sql history as a very usefull feature. One day, I noticed the SQl history does not log any longer any new SQL commands. To resolve this problem, I tried it with a clear history. From that time on the history is empty and the button disabled. What went wrong? I already checked the preferences.
    Client: WinXP prof., SQL developer 1.0.0.15.57, connected Db from 8.x to 10.x

    Seeing as sqlplus worksheet has used F5 for years and saves history, actually, it kind of does. I don't understand why they would intentionally counter the behavior of the long-utilized tool this functionality replaces (and that some of us still have to use on a regular basis). It seems like F9 and F5 should be swapped.
    My 2 cents.
    FYI, I found this thread because I was having the same problem. :)

  • Certain Pages do not show up in my history log. Why is that. I have a security concern as on site was my online bank site.

    I have noticed that some websites will show up in the history log and some website will not. I have a concern that there might be a security breach on my laptop that I am not aware of.

    You probably need to "Publish All" to make all of the
    pages republish and fix their nav bars.
    Hi Allyson
    I did as you said but oh oh.......... it didn't work. I have also been told that I must add the name of the second site to my first site, so that it would look like this:
    http://web.mac.com/lorna6/secondsitename
    I am told to then use this URL and make it a hyper-link to my first site and then my sites will be linked. I take it to mean that when I hyper link it to my first site, this new URL has to be in the Navigation bar at the top of the page.
    Can I name the second site anything I want or am I constrained as I was in the naming of my first site? (the .Mac constraint of having us all use our .Mac name in the URL. BTW, I don't mind this at all, and in fact I think it's a good thing, because that way I can spot another .Mac person.
    Lorna in Southern California

  • Crystal Report 2008 results "Password did not match error" in SQL Server Log

    Hi,
    I am trying to develop some crystal reports using Crystal Report 2008 SP4 to connect to SQL server 2008 through RDO (ODBC). My problem is that anytime Crystal Report is trying to establish a connection with SQL Server, first it tries to connect with the wrong password and results the following error message in SQL Server log:
    Login failed for user 'peyman'. Reason: Password did not match that for the login provided. [CLIENT: 192.168.2.198]
    The login name 'peyman' is the right one as I have quoted the same in setting up ODBC System DSN using "SQL Server Native Client 10". But somehow Crystal Report is taking its chance and trying to connect before prompting me for the password. After this error logged to SQL server side, I can see Crystal Report pops up the prompt to enter DSN password. After supplying password to this prompt, Crystal Report works fine and pulls the data and renders the report without having any more incorrect password error logged to SQL Server.
    The attached file shows the step I am taking to regenerate the issue. I need this to be fixed as anytime uses any of these reports and tries to render it with crystal runtime engine the same error message raises in SQL Server side and logged in the log file.
    Thanks,

    Hi Peyman,
    This is the way it should work, In CR designer when you open the report it does nothing. As soon as you hit the Refresh button then CR tries to use the connection saved with the report. It assumes what is saved with the report is a valid server and connection info. CR simply tries to verify the server is still active.
    If you don't want it to fail in CRD then before opening any report Click the File, Log on Database option and connect. Now when refreshing reports it won't log the failed.
    In the SDK it does the same thing, it assumes the Server info is the same and does try to connect to verify the Server does exist. It only takes a few milli-seconds to do this
    To stop this from happening set the log on info first then it won't log the attempt to connect, it's checking to see if Trusted is allowed.
    TIMESTAMP    THREAD_ID    FILENAME    LINE_NUMBER    LOGGED_DATA    LEVEL
    2014-6-2 8:30:11.439    57320    .\QESession.cpp    444    Set Product View Locale: 4105    20
    2014-6-2 8:30:11.439    57320    .\QESession.cpp    478    Set Preferred View Locale: 4105    20
    2014-6-2 8:30:11.439    57320    .\QESession.cpp    500    Set Process Locale: 4105    20
    2014-6-2 8:30:11.440    57320    .\qecommon.cpp    117    This property is currently in a read-only state and cannot be modified. File Name: ".\QEProperty.cpp". Line: 217    1
    2014-6-2 8:30:11.967    57208    .\QESession.cpp    444    Set Product View Locale: 4105    20
    2014-6-2 8:30:11.967    57208    .\QESession.cpp    478    Set Preferred View Locale: 4105    20
    2014-6-2 8:30:11.967    57208    .\QESession.cpp    500    Set Process Locale: 4105    20
    2014-6-2 8:30:11.968    57208    .\qecommon.cpp    117    This property is currently in a read-only state and cannot be modified. File Name: ".\QEProperty.cpp". Line: 217    1
    2014-6-2 8:30:11.999    57208    .\QESession.cpp    444    Set Product View Locale: 1033    20
    2014-6-2 8:30:12.4    57320    .\qecommon.cpp    117    This value is write-only. File Name: ".\QEProperty.cpp". Line: 145    1
    2014-6-2 8:30:56.278    57208    .\odbcapi.cpp    301    Beginning COdbcapi::DriverConnect    20
    2014-6-2 8:30:56.342    57208    .\odbcapi.cpp    335    Ending COdbcapi::DriverConnect    20
    2014-6-2 8:30:56.342    57208    .\connect.cpp    2170    SQLDriverConnect succeeded: DSN = 192.168.13.172, User ID = sa, Password = ********    10
    2014-6-2 8:30:56.348    57208    .\qecommon.cpp    117     File Name: ".\QEQueryInfo.cpp". Line: 826    1
    2014-6-2 8:30:56.348    57208    .\qecommon.cpp    117     File Name: ".\QEQueryInfo.cpp". Line: 854    1
    2014-6-2 8:30:56.348    57208    .\qecommon.cpp    117     File Name: ".\QEQueryInfo.cpp". Line: 826    1
    2014-6-2 8:30:56.348    57208    .\qecommon.cpp    117     File Name: ".\QEQueryInfo.cpp". Line: 826    1
    2014-6-2 8:30:56.348    57208    .\qecommon.cpp    117     File Name: ".\QEQueryInfo.cpp". Line: 854    1
    2014-6-2 8:30:56.348    57208    .\qecommon.cpp    117     File Name: ".\QEQueryInfo.cpp". Line: 826    1
    2014-6-2 8:30:56.348    57208    .\qecommon.cpp    117     File Name: ".\QEQueryInfo.cpp". Line: 854    1
    2014-6-2 8:30:56.348    57208    .\qecommon.cpp    117     File Name: ".\QEQueryInfo.cpp". Line: 916    1
    2014-6-2 8:30:56.348    57208    .\QERowset.cpp    1184    Beginning CQERowset::readFirstRecord    20
    2014-6-2 8:30:56.348    57208    .\QERowset.cpp    2149    Beginning CQERowset::restart    20
    2014-6-2 8:30:56.348    57208    .\QERowset.cpp    2370    Beginning CQERowset::execute    20
    2014-6-2 8:30:56.353    57208    .\DbQueryBuilder.cpp    514    Query Targets: sqlncli10, ODBC3SQLServer    10
    2014-6-2 8:30:56.353    57208    .\DbQueryBuilder.cpp    525    Successfully built query:    SELECT "Orders"."Customer ID", "Orders"."Employee ID"   FROM   "xtreme"."dbo"."Orders" "Orders"    10
    2014-6-2 8:30:56.353    57208    .\odbcapi.cpp    875    Beginning COdbcapi::ExecDirect    20
    2014-6-2 8:30:56.354    57208    .\odbcapi.cpp    884    Finishing COdbcapi::ExecDirect    20
    2014-6-2 8:30:56.354    57208    .\rowset.cpp    220    SQLExecDirect succeeded:  SELECT "Orders"."Customer ID", "Orders"."Employee ID" FROM   "xtreme"."dbo"."Orders" "Orders"    10
    2014-6-2 8:30:56.354    57208    .\QERowset.cpp    2814    bindToField succeeded: Orders.Customer ID is using client buffer    10
    2014-6-2 8:30:56.354    57208    .\QERowset.cpp    2814    bindToField succeeded: Orders.Employee ID is using client buffer    10
    Notice it doesn't try to connect first if I set the log on info first using code.
    So nothing we can do to stop SQL server from logging this info. Check with the DBA, possibly they can "filter" out the application attempts to connect or change your work flow in the app.
    If you have a Preview Button to view the report then simply add your Database log on info prompt there if Connectioninfo.Trusted is not true:
    mainSecureDB = rpt.Database.Tables[tableIndex].LogOnInfo.ConnectionInfo.IntegratedSecurity;
    if mainSecureDB = false then prompt the user for log on info and set accordingly, if it is true the it should not fail when it connects.
    This is simply a matter of changing your App work flow...
    Thanks
    Don

  • Apply exakt steps from History log file to another image?

    Alright. If you edit an image and the history log option is turned on, you can look at the exakt steps you've made with all the settings for the individual tools and effects used. Also steps you make can be recorded with the actions tool.
    My question is, are there any way to "play" the steps listed in the history log file, as if it were an action?
    I need to apply alot of steps to several images but I never recorded the steps I've made with the actions tool. However they are "recorded" in the history log file and I could go in and look and recreate all those steps manually, but that would take too much time...
    Are there any plug-ins or are there any other way to do what I need to get done?
    Cheers
    Martin

    1. you can set your preferences to record the history to a txt file or the files metadata.
    I assumed that is what the OP already did as indicated in
    If you edit an image and the history log option is turned on,
    wonder why there are no feature in Photoshop to replay all one have done, to any image?
    As you know there is, it’s Actions (as with Scripting not all functionality records; Edit: and not all functionality is applicable to all files because of Color Mode and Bit Depth for example). But it is up to you to employ the necessary foresight to know what to record beforehand.
    Also a non-destructive workflow can help make many effects/compositions/etc. easily transfer-/recreateable. 

  • Quantity of SQL History shrinking

    I've noticed since about Dec. 2006 or Jan. 2007 that the quantity of SQL History kept has varied greatly and generally shrunk. In the last five days it has shrunk from 21 days to 10 days. I currently only have 47 entries in SQL History. I estimate that I execute a minimum of 20 unique queries every business day, so a significant number of my queries over the last 10 days are missing from SQL History.
    In addition, today I'm noticing that I only have entries from a single database connection in SQL History. I have executed queries against two database connections today and at least 4 others over the last week. But only one of those database connections is reflected.
    I've counted the entries in "SqlHistory.xml" and it matches the number of entries I see in the SQL History dialog.
    Any ideas on what is happening to my SQL History? I am aware of the bug that can wipe out all of the SQL History prior to the time when a second instance of SQL Developer is opened. In this case though, I still have 10 days of history and history for all but one database connection is missing.
    SQL Developer 1.1.2.25
    Oracle 10g Enterprise Edition 10.2.0.2.0 - 64bit (all databases I access are the same version)
    Java 1.5.0_05
    Windows XP Pro SP2
    132gb free space on the drive which has both Windows installed and SQL Developer.
    Pentium 4 3GHz
    2GB Ram

    Bill,
    I assume this happened on htmldb.oracle.com. I bet one of our (necessary) cleanup jobs kicked in and purged some old statements.
    Sergio

  • Summary SQL*LDR log file

    Hi,
    I need to get only a summary log file that show me records by error. for example if the log file has 5000 error for datatype and 300 for PK violation and 2000 for not null value, the problem that the log will show me this error row by row. can I get it summary like this
    -5000 rows rejected due to error for datatype
    -300 rows rejected due to PK violation
    -2000 rows rejected due to not null value
    Database: 9i
    O/S: Windows 2000 Server

    eng. Habeeli wrote:
    schavali wrote:
    Pl see your duplicate post here - big SQL*LDR log file
    Pl post what you have found in the documentation so far. AFAIK, there is no way to get a summary log file.
    Sriniits not duplicate ,I change to this form that specialized in sql loadre.
    I don't found any thing about summary log in the documnetPl post a link to the document you read. Did you see my comment above ?
    HTH
    Srini

  • Sql server log shipping space consumption

    I have implemented sql server logs shipping from hq to dr server
    The secondary databases are in standby mode .
    The issue is that after configuring it , my dr server is running out of space very rapidly
    I have checked the log shipping folders where the trn files resides and they are of very decent size , and the retention is configured for twenty four hours
    I checked the secondary databases and their size is exactly same as that of the corresponding primary databases
    Could you guys please help me out in identifying the reason behind this odd space increase
    I would be grateful if you could point me to some online resources that explains this matter with depth

    The retention is happening . I have checked the folders they do not have records older then 24 hours .
    I dont know may be its because in the secondary server (Dr) there is no full backup job working , is it because of this the ldf file is getting bigger and bigger but again I far as my understanding goes we cannot take a database full backup in the stand
    by mode .
    The TLog files of log shipped DBs on the secondary will be same size as that of primary. The only way to shrink the TLog files on secondary (I am not advising you to do this) is to shrink them on the primary, force-start the TLog backup job, then the copy
    job, then the restore job on the secondary, which will sync  the size of the TLog file on the secondary. 
    If you have allocated the same sized disk on both primary and secondary for TLog files, then check if the Volume Shadow Copy is consuming the space on the secondary
    Satish Kartan www.sqlfood.com

  • SQL History - returning wrong statement if history is sorted

    I have searched the forum and not seen anything on this, but I have been getting the wrong SQL statement returned when I select Append/Replace.
    I have statements going back to 12-Jan-06 (using 804 since 13-Jan-06, 796 since 12-Jan-06 and 715 last year), so I don't know if I have hit the aging limit or not (I think I counted 221 entries).
    It seems to me that it works out the cursor position and then returns that statement number from the history - independently of resorting the history.
    For example, if I sort from oldest to newest (based on timestamp) and then select the bottom statement (displays as newest) and Replace, I get the oldest statement in the worksheet. If I select the top statement (displays as oldest) and Replace, I get the newest statement in the worksheet.
    This is further complicated by the fact that sorting on the Timestamp is doing an alpha-numeric sort, rather than a date sort. When sorting from oldest to newest, the bottom three timestamps are (in "oldest to newest" order) are:
    1/23/06 10:58 AM
    1/23/06 9:21 AM
    1/23/06 9:31 AM
    If I then remove the sorting, the statement returned is the correct one.

    It was great to get the fix for the SQL History returning the right statement if sorted (we got it in v1184).
    However v1215 still does a text sort on the TimeStamp column rather than a date sort, so 2PM comes before 9AM, etc.
    Also, where is the format for the TimeStamp column coming from? As someone who doesn't live in the US and thinks that days should come before months in date formats (ie DD-Mon-YYYY), I am frustrated by yet another product that uses MM/DD/YY, despite my best efforts at telling it otherwise.

  • Activation of History Log for Material Classification CLASS TYPE 001

    Hi All.
    I've a question about this object.
    We need to launch RCCLUKA2 to activate multiple objects registration and to generate history log for changes in classification of materials (AUSP table).
    Reference are sap notes 942691 and 943559.
    I FOUND OUT THERE'S a big problem. RCCLUKA2 change AUSP objek key, using internal number range from INOB objects.
    There're hundreds of custom reports and other stuff that read AUSP directly with Matnr Code as objek key.
    I think that's a serious problem, isn't it? Or is there any solution about ?
    Thanks for Help!
    _______> old message
    This report RCCLUKA2 does 5 steps.
    1 step: activation flag for multiple objects in table TCLA and TCLAO only for  input class type (in this case 001 -> MATERIAL);
    2 step: download of KSSK data and AUSP data in two text files (name defined in input parameters);
    3 step: delete all KSSK data and AUSP data for class type 001,  10.000 records every loop;
    4 step: create reference object in table INOB for all records inserted in KSSKFILE, inserting all records in KSSK table at END of INOB upgrade;
    5 step: inserting record in AUSP table from all records inserted in AUSPFILE, using INOB objectid created in step 4.
    DEQUEUE ALL
    We think that this process is a little bit dangerous in production system, due to procedures that always run, in background or foreground.
    I've found a work around that creates an implicit enhancement for LCLVFF10 form FILL_REDUN.
    I think that it's not suggested to create this enhancement because of the missing flags in TCLA/TCLAO and missing reference in INOB and KSSK database.
    This enhancement would create records only in CDHDR and CDPOS with a general key object reference (such concatanate matnr(without-letters) mafid and atinn code).
    This enhancement couldn't also activate the standard function of reading document changes from standard menu in MM03/MM02/MM01, etc..
    Do you have any suggestion about this?
    RCCLUKA2 is a very old report. SAP didn't give any advice or upgrades for this need?
    Thanks for your help
    Message was edited by: Roberto Vacca

    Dear ,
    changes in values for a characteristic can be monitored in CL20N but as for MM02/MM03 transaction only for class type 023.
    For class type 001 or 300 the change document isnu2019t available.
    Thanks.

Maybe you are looking for

  • Error while scheduling a Hyperion Interactive Report

    Hi, I'm trying to run a scheduled Hyperion Interactive Report (version 9.3.1). The report is scheduled to run for 1st of every month. After the report is scheduled. When a user tries to run the report on demand. The error log states as follows: $ vie

  • Crystal report XI error while Linking tables

    We have just purchased Crystal Report XI  license copy  but as we going to create new report there is problem in linking tables i.e while pressing LINK tab after selecting ODBC databse  name  application going shut down.Please suggest.Thanks in advan

  • Error saving column settings with ORA-01403: no data found Error

    Hi I have a region with multi row report updatable. When I am trying to change the query I am getting below error. Anybody let me know what will be the problem. ORA-01403: no data found Error saving column settings Regards Kiran

  • Disk Utility from external drive?

    Our mac mini has a corrupted disk and needs Disk Repair, but the DVD drive is also broken.  Can I boot from an external DVD drive?

  • Snapshot naming

    Snapshot naming: - Default to date, date/time or last edit name (default naming selectable via preferences)