Sequence Behavior - not clear

Hi,
I cannot understand the behavior of this sequence
CREATE SEQUENCE NS5;
ALTER SEQUENCE NS5
INCREMENT BY 10;
SELECT NS5.NEXTVAL FROM DUAL;
Why is the first value being generated : 10
The default start with value is 1, why isn't 1 displayed.
Thanks for your responses in advance.

>
Why is the first value being generated : 10
The default start with value is 1, why isn't 1 displayed.
>
That is by design - SEQUENCEs are documented to produce the behaviour that you experienced.
See ALTER SEQUENCE in the SQL Language doc
http://docs.oracle.com/cd/E14072_01/server.112/e10592/statements_2012.htm
>
•If you change the INCREMENT BY value before the first invocation of NEXTVAL, then some sequence numbers will be skipped. Therefore, if you want to retain the original START WITH value, you must drop the sequence and re-create it with the original START WITH value and the new INCREMENT BY value.

Similar Messages

  • Payment not  clearing for Vendor through Automatic Payment Program

    Hi
    I had setup automatic payment program for my vendor to take $ from my GL account. When I run the payment program, it ran without any error, however, the vendor account was not cleared. The balance in GL and Vendor account remains the original amounts.
    The log shows the following.
    Job started
    Step 001 started ( program SAPF110S, variant &0000000001081, user ID XXXXXXX)
    Log for payment run for payment on 01/12/2008, identification XXXX
    End of log
    Job finished
    Can anyone help ?
    Thanks
    Shalini

    I have not proposed a solution. Now go to F110 and go to the table additional tab and press proposal log button on the lower right hand side. this will give a detailed log. Either analyse this error set right error or send the error log for analysis.
    First run the proposal run and if it is successful, run the payment run and then to print run.
    It should be run in the above sequence.
    Edited by: Sivakumar Gopalakrishnan on Jan 12, 2008 7:54 PM

  • Sequence will not validate with run time prompts in business rules

    I am in Hyperion Planning v 1.1.1.3, with Workforce initialized. I am creating a sequence of business rules. The rules use run time prompts within them and validate when used on their own. However, when I insert them in a sequence the sequence will not validate, apparently due to the run time prompts.
    The sequence returns the error: Cannot calculate. Essbase Error(1200323): Error parsing formula for [FIX STATEMENT] (line 7): expression expected after [(].
    Line 7 in the business rule is as follows:
    FIX ([rtpYear],[rtpScenario],[rtpVer],[rtpCC],[rtpBU],[rtpEmp])
    The problem is not isolated to this business rule, it occurs with any rule I try.
    Can sequences function with run time prompts in the included business rules?
    Stephen

    I have figured out that if any rtp is hidden in the business rule, the sequence will not validate. Is this expected behavior for sequences?
    Additionally, hiding the variables in the sequence has no effect if the variables are not hidden in the rule. When launched from a menu in a web form, the user is prompted for values for all variables (except for the Execution Database Name) although they are prepopulated based on the members in the POV, Page, etc. Is this also expected behavior?
    I also find that it won't validate if "Merge launch variables" is selected (I have included rules associated with different plan types and with different dimensionality, so this may be expected). The user is prompted to provide values for every variable in each rule in the sequence, many of them duplicates (though, again, they are prepopulated based on members in the form).
    Stephen

  • Java persistence: sequence generator not working in expected way.

    Hi,
    I just played around with java persistence api and found that sequence generator annotations for making new primary key value not working correctly in my code. following is part of my code related to this problem.
    in entity class:
    @SequenceGenerator(name="mySeq", sequenceName="MY_SEQ_IN_DB")
    @Id
    @GeneratedValue(strategy=SEQUENCE, generator="mySeq")
    @Column(name = "UPLOAD_ID", nullable = false)
    private Integer uploadId;
    I have a sequence called MY_SEQ_IN_DB in my oracle 9i testing db. When I ran the code to save a new entity instance to database I found that the new upload id is not the one that should be created from the MY_SEQ_IN_DB sequence, while the sequence was not increased either. It's instead something starting from 97 (in my pc), and increased just like a sequence, but every time I reran my code the sequence restarted itself again beginning with 97(of course the insert failed with primary key conflict). the behavior just looks like it's a sequence created and managed totally by the persistence api or toplink itself (I guess).
    So, my question is, is this a designed behavior or have I done something wrong to make it not using the correct sequence in the db?
    I used hibernate before, and assigning new primary key from sequence in this way works fine.
    Thanks
    JC.L

    There are some things in Java which can't be serialized. When contained within a higher-level class such as Buttons. A button can be associated with an Action Listener. Now, you can serialize a button; however you can't serialize an action listener - for what should be to you obvious reasons. This is NOT an error and so there is no need to throw an exception when serializing the button.
    Basically: Applet sends servlet button, but not listener.

  • After hiighlighting a word and pasting it into another document, the clipboard does not clear ;when I hit enter it pastes again and again. How do I fix it?

    When I highlight a word or phrase and copy to the clipboard, then paste it in another area the clipboard does not clear the phrase. Therefore, every time as I type and hit the enter key it continues to paste the phrase all over my document. I do have the latest download of Firefox installed (16.0, win 7 and that is when this started. How do I stop this obnoxious behavior so I can type a sentence with normalcy?
    Louise

    Thank you, Iusually do that first thing with any computer problems. Old school computer fix. Hard Boot regardless! lol Thanks again.

  • Vendor Advance not cleared

    Hello Sir,
    I am doing vendor down payment t code:f-48 and clear from t code:F-54
    But entry shown in FBL1n with tick Spl.G/L ind.in credit balance ? its shows Red indicator?why not in Green indicator ?
    Waiting for Reply.
    Thanks
    J

    Dear J,
    what You' re reporting is the normal System behavior.                                                                               
    When a down payment is cleared with transaction F-54, the down payment                                     
    clearing document contains 2 vendor line items: one special G/L line                                       
    item and one 'credit memo' line item. The special G/L line item clears                                     
    the special G/L line item in the downpayment, whereas the 'credit memo'                                    
    line item remains open and does not clear the vendor line item in the                                      
    invoice.                                                                               
    But this is correct: Usually the amount of the vendor line item                                            
    in the invoice (-> the amount of the 'final' invoice) is higher than the                                   
    amount in the 'credit memo' line item of the downpayment clearing (the                                     
    amount of the 'credit memo' line item of the downpayment clearing is the                                   
    amount of the downpayment), and so these two vendor line items cannot be                                   
    cleared with each other. But if they should, the customer could clear                                      
    them and post a residual line item with another clearing transaction.                                                                               
    If you want to get another result, like clearing the documents, you have                                   
    2 methods.                                                                               
    1) using F-53 (outgoing payment post) to clear the invoice and the                                         
    downpayment.                                                                               
    2) after using F-54, you have to run F-44 (account clear) to fully clear                                   
    the open item.      
    I hope this helps You.
    Mauri

  • Cannot allocate new log, sequence Checkpoint not complete

    Hi,
    Im having very frequent log switches and Im getting error as
    " cannot allocate new log, sequence Checkpoint not complete"
    I was having 3 redo log groups with 50 MB each. After I found this error in the alert log; I increased teh number of redo log groups to 6 with 50MB each. Still the issue is not getting resolved.
    Please suggest what will be the best solution for this.
    Following is a snippet from alertlog.
    ==================================================
    Sun Apr 19 09:14:08 2009
    Thread 1 advanced to log sequence 5811
    Current log# 2 seq# 5811 mem# 0: /u01/app/oracle/oradata/mview/redo02.log
    Thread 1 cannot allocate new log, sequence 5812
    Checkpoint not complete
    Current log# 2 seq# 5811 mem# 0: /u01/app/oracle/oradata/mview/redo02.log
    Sun Apr 19 09:14:18 2009
    Thread 1 advanced to log sequence 5812
    Current log# 3 seq# 5812 mem# 0: /u01/app/oracle/oradata/mview/redo03.log
    Thread 1 cannot allocate new log, sequence 5813
    Checkpoint not complete
    Current log# 3 seq# 5812 mem# 0: /u01/app/oracle/oradata/mview/redo03.log
    Thread 1 advanced to log sequence 5813
    Current log# 1 seq# 5813 mem# 0: /u01/app/oracle/oradata/mview/redo01.log
    Thread 1 cannot allocate new log, sequence 5814
    Checkpoint not complete
    Current log# 1 seq# 5813 mem# 0: /u01/app/oracle/oradata/mview/redo01.log
    Sun Apr 19 09:14:32 2009
    Thread 1 advanced to log sequence 5814
    Current log# 2 seq# 5814 mem# 0: /u01/app/oracle/oradata/mview/redo02.log
    Thread 1 cannot allocate new log, sequence 5815
    Checkpoint not complete
    Current log# 2 seq# 5814 mem# 0: /u01/app/oracle/oradata/mview/redo02.log
    Thread 1 advanced to log sequence 5815
    Current log# 3 seq# 5815 mem# 0: /u01/app/oracle/oradata/mview/redo03.log
    Thread 1 cannot allocate new log, sequence 5816
    Checkpoint not complete
    Current log# 3 seq# 5815 mem# 0: /u01/app/oracle/oradata/mview/redo03.log
    Sun Apr 19 09:14:44 2009
    =========================================================
    Regards
    Pratheej

    Anand... wrote:
    Hi Sir,
    Although i too had suggested increasing the redo logfile size, but after going through [http://asktom.oracle.com/pls/asktom/f?p=100:11:0::::P11_QUESTION_ID:69012348056] i was little confused as Tom has mentioned
    Another way is to make the log files smaller, hence increasing the frequency with which we checkpoint ---- Can you explain why so.
    I was a little surprised when I read that posting - but noticed it was dated May 2000 - when databases were still quite small and less busy. (And Tom suggested 25MB as being "modest" rather than "tiny" - which is probably what many people would call 25MB these days). And May 2000 probably means 8.0 or 8.1 - and the whole log buffer, redo generation, checkpointing technology has changed a lot since then.
    Basically, if you hit "checkpoint not complete", you need more online log space so that it's possible during the busiest times to keep generating redo log information while the checkpoint queues are being cleared far enough to allow older log files to be recycled.
    You can do this by adding more log files, or by increasing the sizes of the log files you use. Tom's point, I think, was that if you chose the option to add more files and kept them small (or even made them smaller) then the volume of dirty data blocks that you could create while filling a log file would be small, so the database writer wouldn't have to do much work to make each log file available for re-use. (I'm not sure I'd agree with the approach, though - even for 8i - because it could easily lead to an increase in the volume of datablocks written, even if it did bypass the checkpoint issue).
    Regards
    Jonathan Lewis
    http://jonathanlewis.wordpress.com
    http://www.jlcomp.demon.co.uk
    "For every expert there is an equal and opposite expert."
    Arthur C. Clarke

  • Sequence behavior after importing via DataPump

    Hi Friends,
    I'm running Oracle DB 11.2.0.3 on Windows 2008 R2 SP1 Servers and I faced a strange sequences behavior after importing a schema via Data Pump.
    The export is done this way:
    EXPDP userid/password dumpfile= logfile= directory= remap_dumpfile=y (no news)
    The import is done this way
    IMPDP userid/password dumpfile= logfile= directory= remap_schema=(old_one:new_one) remap_tablespace=(old_ones:new_ones, so on...)
    The import works fine. There are no errors and the sequences are as well imported with no warnings.
    The strange behavior is that the sequences seems to "reset". When we call a sequence the NEXTVAL is just lower than the values already stored in the Database, and we get ORA-00001 a lot. The sequence should know that vale. I don't have this problem when using exp/imp, just via DataPump.
    So that when we create an order that should receive the value of 100, as an example, because we have 99 orders on the system, Oracle suggest a value lower than 99 or even the number one value (01).
    We then wrote a script to check the CURVAL of the sequences on the base schema to recreate the sequences using this initial value on the new imported schema.
    Does anyone faced this problem before?
    Any suggestions?
    Tks a lot

    Richard
    I've tried what you just said.
    Adding the parameter consistent=y makes Oracle to show a message like that at the beginning of the export
    "flashback_time=TO_TIMESTAMP('2013-09-03 12:18:12', 'YYYY-MM-DD HH24:MI:SS')"
    It warns me: Legacy Parameter CONSISTENT=TRUE, and replaces with flashback_time.
    Really, I did not know about this behavior with this "old" parameter. I'm very appreciated about your help.
    I was almost thinking it was a DataPump bug or something.
    Thanks a lot Richard. I'll now update my scripts and make lots of test.
    If you have more advices using this parameter please share us.
    Cheers

  • UCCX Real time Reporting - Engine Tasks not clearing

    I've got a couple of VERY basic test scripts that I wrote for testing session management from one script to the next.  I finally got the scripts to work, but I've been getting this error message when the first script completes (after the call is transferred to the second script):
    Contact ID: <xxxxx>, Contact is inactive when getting channel.
    I didn't really think too much of it until I decided to look in RTR to make sure nothing was getting hung.  Turns out, when I view the "Engine Tasks", one task exists for each time I've called each script.  There are 11 tasks right now and periodically other tasks will pop in and out from our production scripts, as they are supposed to.  I'm assuming I will have to restart the engine service to clear the tasks (although letting it go overnight seems to have cleared the tasks from yesterday), but what am I doing wrong in the script that is not clearing the tasks? 
    For moving the call from the first script to the second script, I am using a "Call Redirect".  In the "Successful" branch, it goes through a "Set Contact Info" and marks the call as handled, then a "Go to Terminate" step, which is where the error is generated the script ends. It never actually makes it to the "End" step.
    Thanks
    Chris

    I made one change to my script that was able to get rid of the error message, but the Engine Tasks are still there.  I tested this morning and just checked; they are still going at 4.5 hours.
    I've attached a screenshot of the first script.  As you can see, it is very basic, but I feel like I am missing something.  I didn't post the second script, but it is very similar, except no call redirect.  It has the same behavior in that it still creates an Engine Task that takes many hours to clear.
    Thanks for taking a look!

  • ORA-16014: log 2 sequence# 283 not archived, no available destinations

    When trying to start my database my alert log is showing
    ORA-16014: log 2 sequence# 283 not archived, no available destinations
    ORA-00312: online log 2 thread 1: '/home/app/oracle/oradata/PRIMARYP/redo02.log'
    I can mount the instance ok but not actually open the database
    my db_recovery_file_dest_size is 22 Gb and and the mount point defined by db_recovery_file_dest has pleanty of free space
    Also
    select *
    from v$recovery_file_dest;
    shows plenty of free FRA space -
    NAME
    SPACE_LIMIT SPACE_USED SPACE_RECLAIMABLE NUMBER_OF_FILES
    /home/app/oracle/fast_recovery_area
    2.3622E+10 5506131968        3297617408             159
    select error
    from v$archive_dest;
    shows no errors
    Can anyone help !
    thanks, Jim

    ok here's the wierd thing. When I set up the primary, physical standby and logical standby, I set the following Log Archive Destinations on my primary database ( into the SPFILE )
    ALTER SYSTEM SET log_archive_dest_1='LOCATION=USE_DB_RECOVERY_FILE_DEST VALID_FOR=(ONLINE_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=PRIMARYP' SCOPE=BOTH;
    ALTER SYSTEM SET log_archive_dest_2='SERVICE=STANDBYP ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=STANDBYP' SCOPE=BOTH;
    ALTER SYSTEM SET log_archive_dest_3='SERVICE=STANDBYL ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=STANDBYL'
    SCOPE=BOTH;
    ALTER SYSTEM SET log_archive_dest_4='LOCATION=USE_DB_RECOVERY_FILE_DEST VALID_FOR=(STANDBY_LOGFILES,STANDBY_ROLE) DB_UNIQUE_NAME=PRIMARYP' SCOPE=BOTH;
    Then I converted to using DGMGRL instead od SqlPlus to manage my Data Guard Environment. I then did a switchover to my Physical Standby and then a switchbach to my primary. Today when I check the log_archive_dests on my primary, they are completely different from how I set them up originally !
    *.log_archive_dest_1='service="STANDBYL"','LGWR ASYNC NOAFFIRM delay=0 optional compression=disable max_failure=0 max_connections=1 reopen=300 db_unique_name="STANDBYL" net_timeout=30','valid_for=(all_logfiles,primary_role)'
    *.log_archive_dest_2='service="STANDBYP"','LGWR ASYNC NOAFFIRM delay=0 optional compression=disable max_failure=0 max_connections=1 reopen=300 db_unique_name="STANDBYP" net_timeout=30','valid_for=(all_logfiles,primary_role)'
    *.log_archive_dest_3=''
    *.log_archive_dest_4='LOCATION=USE_DB_RECOVERY_FILE_DEST VALID_FOR=(STANDBY_LOGFILES,STANDBY_ROLE) DB_UNIQUE_NAME=PRIMARYP'
    As you can see there is no setting place for the standard archive logs for PRIMARYP !
    When I set log_archive_dest_3 to
    'LOCATION=USE_DB_RECOVERY_FILE_DEST VALID_FOR=(STANDBY_LOGFILES,STANDBY_ROLE) DB_UNIQUE_NAME=PRIMARYP' SCOPE=BOTH;
    Then I am able to open the PRIMARYP database again.
    Q. what cause the entire resetting of the 4 log_archive_dest_n settings ? I am presuming it is something the broker has done ( presumably on the switchover or switch back ) ? However it seems to have completely mucked itself up, not even provising a standard archive log destination for PRIMARYP and in the process causing the instance not to be able to start
    Has anyone ever seen this before ?
    Also now I am getting  ORA-16783: cannot resolve gap for database STANDBYP, STANDBYL
    In my Broker Config for the PRIMARYP database. I am worried because during the investigation of this problem I issues the following on PRIMARYP
    alter database clear unarchived logfile group 1;
    Have I created a Gap which I cannot resolve ?
    thanks,
    Jim

  • CS4 - does not clear session info

    I developed a small membership site for my club.
    Everything works fine so far. I can log in and see the protected pages. I can log out and try and see the protected pages but I then get redirected to the page I specified for failed login.
    This shows me that the login and logout function works.
    The Problem
    The problem is I logged in as me and I log out again. I then logged in as a different user and tried to update the different user's details. But when I went to the update form it still showed MY details in the update form and not those of the user who is currently logged in.
    So it seems to me as if CS4 uses session variables to remember the login details but it was not cleared after the first users logged out.
    Any suggestions on how to fix this?

    pietpetoors wrote:
    The problem is I logged in as me and I log out again. I then logged in as a different user and tried to update the different user's details. But when I went to the update form it still showed MY details in the update form and not those of the user who is currently logged in.
    So it seems to me as if CS4 uses session variables to remember the login details but it was not cleared after the first users logged out.
    The code generated by the Dreamweaver Log Out User server behavior looks like this:
    if ((isset($_GET['doLogout'])) &&($_GET['doLogout']=="true")){
      //to fully log out a visitor we need to clear the session varialbles
      $_SESSION['MM_Username'] = NULL;
      $_SESSION['MM_UserGroup'] = NULL;
      $_SESSION['PrevUrl'] = NULL;
      unset($_SESSION['MM_Username']);
      unset($_SESSION['MM_UserGroup']);
      unset($_SESSION['PrevUrl']);
      $logoutGoTo = "login.php";
      if ($logoutGoTo) {
        header("Location: $logoutGoTo");
        exit;
    As you can see, it's removing only the session variables created by Dreamweaver.
    A more secure way of handling this would be to change the code to this:
    if ((isset($_GET['doLogout'])) &&($_GET['doLogout']=="true")){
      //to fully log out a visitor we need to clear the session varialbles
      $_SESSION = array();
      session_destroy();
      $logoutGoTo = "login.php";
      if ($logoutGoTo) {
        header("Location: $logoutGoTo");
        exit;

  • FDM export to HFM - 2nd export did not clear all data from 1st export?

    Has anyone experienced exporting a replacement FDM file to HFM where the 2nd file did not contain an account that was present in the original file and that account/value in the 1st file then remained in the HFM trial balance and caused it to be out of balance? Does the replace export option not clear all the prior file data? The only unusual part of it might be that the problem account was derived using logic group script? Thanks.

    Thanks Hensen - Sure, it was easy enough to fix once I deduced what the problem was - although my solution was just to use the data clear function and reload. The real question is how to prevent it from happening again and it sounds like I can't do much about it - other than trying to train the users to watch out for do-over files that have fewer accounts than the original load?? Am I wrong to think that HFM should try to fully clear all previous data from a load file if there is a replacement or at least try to validate whether or not all prior accounts have been repopulated?? There's plenty of good validation in the validation step of FDM. Why not during the export step? If this makes sense, I'll put in another "enhancement request" to support.

  • We can not clear all recents(30 recents call).  When Tango show "No recents call". We touch other buttons and come back to Recents button again.  It still show 30 recents call.

    We can not clear all recents(30 recents call).  When Tango show "No recents call". We touch other buttons and come back to Recents button again.  It still show 30 recents call.

    No one here is going to do anything about it. Send feedback to Apple.
    http://www.apple.com/feedback/ipad.html
    Basic troubleshooting steps. 
    17" 2.2GHz i7 Quad-Core MacBook Pro  8G RAM  750G HD + OCZ Vertex 3 SSD Boot HD 

  • System not clearing the Payment request items with the payments

    HI Everybody,
    User getting an issue relating to clearing of the payment request items. In a particular Contract account, it has got 51.75 Pounds of Bank Giro incoming payment. It has contains some 5 items of payment, 3 items of 10.75 and 1 Item of 21.70. The 3 items of 10.75 were cleared 3 payment reqquest items and since there is no due payment requests were there at that time, this 21.70 was showing pending credit item. After two days  two payment requests came as due. There are jobs which run daily for FPMA. But this 21.70 is not clearing to this due payemnt requests. I tried to run this FPMA manually for this contract account. But these were not cleared.
    Could anybody suggest anything on this why these items were not clearing against each other?
    Thanks & Regards
    Siva

    Hi
    The issue is due to the implementation of support packs. Specifically to do with note 1005708.
    siva

  • LM13: Verification field not cleared after [Enter] key

    Hi Experts,
    I'm facing the issue as stated in the OSS note 916654 - LM13: Verification field not cleared after [Enter] key. My current system is ECC, that is why the OSS is already applied to the ECC version. But the problem is, i still faced the problem stated in the OSS.
    Symptom
    In customizing:
    - You have defined [ENTER] in destination screen 302 to SAVE items.
    - You want to verify the destination bin information.
    You use 'Confirm Transfer Order Clustered' RF transaction LM13.
    You scan several SUs and press enter.
    You press [F4-Next pushbutton] to confirm the associated TO(s).
    In the destination screen 302, you scan the verification value for the first item and press [ENTER] key.
    ==> The item is confirmed correctly but when the system proposes the next item, the verification value of the first item is displayed.
    Anyone know how to fix this issue? Pls help.
    thanks
    Hooi Fung

    Hi,
    For example itab-matnr is your screen field
    in PAI
      field: itab-matnr,
             module check_matnr_0700.
    module check_matnr_0700 input.
      move itab_matnr to v_matnr.
    endmodule.
    in PBO.
    move v_matnr to itab-matnr.

Maybe you are looking for

  • Yosemite mail flagged mailbox

    I had my flagged inboxes named according to my preferences in Mail when I was running Maverick. E.g., the red flag had ben renamed "Critical". Now, every time I start mail, the flagged mailbox revert to the original name (e.g., "Red"). If I try to ch

  • Date changes with iphoto 6

    When I send photos from my Samsung D900 to my Mac (via bluetooth) the date info does not seem to come with it. Instead the file is dated as per the date of import onto my computer. When I then put this into iphoto the date changes again to 01/01/06.

  • Master data Load from Oracle table

    Hi, My master data comes from an Oracle table. The table has both attributes and texts combined. How do I load the Master data into the text and attribute info objects thru Direct Upload of Master Data? Is it necessary to go in for Flexible Upload in

  • MDM Search Webdynpro Component

    We have MDM 7.1 SP04 installed. I am using MDM Webdynpro components for MDM 7.1 SP04 running on CE 7.1 EHP1 SP04.  I configured the search component in Webdynpro configurator. My requirement is to pull only active items from MDM for any type of searc

  • Conditional record

    hi experts, Batch Input for Creation and Change of condition records for condition types        with    different access sequences in sales and distribution. what did it mean explaine me briefly thanks in advance.