When should I logout my server session?

Hello Everyone,I am a newbie using TopLink.
I create my server session when my application is loaded, and acquire a client session from this server session for every client operation.However,I don't know when should I call logout method of the server session,as mentioned in Developer's Guide, "When you are finished using a database or server session, you must log out of the session using Session method logout. Although TopLink provides a finalizer to release sessions, this is a last resort. Oracle recommends that you always log out of your sessions."
Many thanks in advance.

You should only logout the ServerSession once your application will no longer needs it. For most applications this is not until the application is shutdown.
--Gordon                                                                                                                                                                                                                                                                                                                                               

Similar Messages

  • How to close a session when user clicks Logout

    In my jsp page i create a session for every user's name.
    when particular user clicks logut without closing his personal page,
    the others can access his page.
    so, how to close a session when clicks logout link.?
    similarly when user gets signin one system,he will not be sigin in other system.but how can we achieve this?
    Thanks in Advance

    What is the name of the session variable that you are creating on login? When he clicks logout, set that session variable to null. Like session.setAttribute("userName",null); and check for non-null condition before proceeding with any functionality in your jsp's.
    What is the second question? You want to prevent duplicate logins from different systems? You would have to maintain a static datastructure like a HashMap for instance. Everytime a user logs in, put an entry into the HashMap...like userMap.put("userName","userName"); If an entry already exists, throw a message saying that user already logged in from another machine...

  • Starting a managed server session and keeping the session when you exit

    I created a managed server named "xxx_mgd" with the create managed server wizard in the admin console
    when I start the managed server it produces a prompt for the username and password, whcih is the same username
    and password entered when i was created. The issue I have is that when I exit the session, the process dies.
    In comparison to the starting of the admin server, which is started with custom scripts which calls the standard wl startup scripts and it runs in background using nohup....
    Is there away to start my managed server the same way, so that it stays running when you exit. I have tried running it as the admin server, using an option where the password if fed, but it dies and never startup
    If you have any clue let me know
    Joseph

    Assuming that you are running WebLogic version 9.x and above.
    Copy the "security" folder inside DOMAIN_HOME/servers/AdminServer directory to DOMAIN_HOME/servers/ManagedServer directory.
    This security folder will have boot.properties file which contains encrypted username and password which is used when starting the server.
    You can run Admin Server in background using nohup command as below from DOMAIN_DIRECTORY. This command will redirect stdout to AdminServer.out file.
    nohup ./startWebLogic>>AdminServer.out &
    You can run Managed Server in background using nohup command as below from DOMAIN_DIRECTORY/bin folder. This command will redirect stdout to ManagedServer.out file.
    nohup ./startManagedWebLogic.sh <Managed Server Name> t3://<AdminServerListenAddress>:<PortNo> >>ManagedServer.out &
    Hope this helps.
    - Tarun
    Edited by: Tarun Boyella on Feb 3, 2011 2:09 PM

  • How to create a dedicated-server session (to a oracle database)

    Hello!
    I've installed Oracle SQL Developer (Rel.: 2.1.1.64.45) and Oracle Instant Client (Rel.: 11.1.0.7), my Client is configured to enable connections using tnsnames.ora or host-/easy-naming (sqlnet.ora-entry: NAMES.DIRECTORY_PATH= (TNSNAMES, HOSTNAME)).
    Regardless how I configure a database-connection in SQL Developer, I allways "only" get a shared-server connection (even when using a tnsnames.ora-entry with "...(SERVER = DEDICATED)...").
    SQL Developer (any release) in combination with ("fat") client 9.2.0.7 (using tnsnames.ora) creates dedicated-server sessions as wanted.
    Is there any setting in SQL Developer which dis- and/or enables dedicated-server sessions or is this a feature/problem of Instant Client 11.1.0.7?
    Any ideas and/or suggestions?
    (Additional Info: There is no difference in this behavior according to the connected database, I work with databases 9.2.0.7, 10.2.0.3, 10.2.0.4, 11.1.0.6 and 11.2.0.1)
    With kind regards
    Klaus

    Hi,
    Take a look at the main [ SQL Developer page on OTN|http://www.oracle.com/technology/products/database/sql_developer/index.html]. There is a link you should find useful, it's the Getting Started link under the Migration section. This explains how to download and setup the drivers you need to correct to non-Oracle databases. There are also links to online demonstrations on how to create connections to these databases. See how you go with that material.
    It should help.
    Sue

  • Multiple server sessions

    Hi,
    Should there be any issue mixing objects between multiple server sessions.
    We have multiple services that each have each have their own ServerSession.
    There are situations where objects retrieved through one ServerSession are ultimately part of an object that is persisted through another ServerSession.
    Note: communication between client and services is via RMI and persistence is performed using the mergeWithReferences method.
    Example,
    - Person object has an Address object.
    - Address is retrieved from ServerSession 1 (RMI)
    - Person object is retrieved from ServerSession 2 (RMI)
    - Address is added to Person and Person is persisted
    thru ServerSession 2. (RMI)
    The behavior I've observed is inconsistent but ultimately there seems to be problems.
    From the example above, occasionally ServerSession2 will attempt to reinsert the Address object into the table.
    In the case where Address is read only the cached version of Person in ServerSession2 will have a null Address attribute after persistence even though the database was updated correctly.
    What accounts for this behavior?
    Thanks
    Mark

    Marc,
    The problem you are seing is related to TopLink's existence checking. When you go to write an object from session 1 into session2 and it does not exist in the cache TopLink assumes the object is new. This is very similar to when you are running in a cluster and your write request ends up in a JVM where the TopLink session has not read in the object you plan to write.
    In order to make this work you can change the existence-checking setting, which I don't recommend, or you can alter your pattern for dealing with the UnitOfWork.
    When you serialize the object(s) across RMI to the server that is going to do your writing I recommend the following pattern:
    1. Acquire UnitOfWork from session
    2. Read objects from database
    unitOfWork.read(person);
    unitOfWork.read(address);
    3. Merge objects (unitOfWork merge APIs)
    4. Commit UnitOfWork
    Step #2 is important here. If you are writing to a session that has the objects cached then no database call is required. If not, the database version will be read in. The merge will copy the values over to the working copy and the commit will calculate and write any necessary changes.
    Other recent threads in this forum and the documentation detail the merge options.
    Doug

  • UnitOfWork and server session cache

    Hello,
    seems like in some situations UoW does not update server session cache:
    SessionManager manager = SessionManager.getManager();
    Server server = (Server) manager.getDefaultSession();
    Session client = server.acquireClientSession();
    UnitOfWork unit = client.acquireUnitOfWork();
    // stored procedure call preserves caching !
    StoredProcedureCall call = new StoredProcedureCall();
    call.setProcedureName("security.set_current_user");
    call.addNamedArgumentValue("p_user", 123);
    unit.executeNonSelectingCall(call);
    // these objects should always be cached (if caching is enabled) !
    List list = unit.readAllObjects(Naudotojas.class);
    unit.commit();
    client.release();
    server.release();
    The presented source code executes stored procedure and read query within singe transaction. Problem is that selected data is not cached. But if stored procedure call is removed, then records will be cached. Is it expected behaviour ? That workarounds could be applied to force caching ?

    When the first modify call is performed in a UnitOfWork it then assumes that the transactional state of the database is transient and therefore and data read could be non-committed. At this point it is not safe to load the objects into the shared cache.
    If you are attempting to make use of VPD security you should read up on the support offered in TopLink. It will give you additional call-backs for setting the user credentials.
    Doug

  • Voicemail access issue when migrated to new Domino server

    st1\:*{behavior:url(#ieooui) }
    /* Style Definitions */
    table.MsoNormalTable
    {mso-style-name:"Table Normal";
    mso-tstyle-rowband-size:0;
    mso-tstyle-colband-size:0;
    mso-style-noshow:yes;
    mso-style-parent:"";
    mso-padding-alt:0in 5.4pt 0in 5.4pt;
    mso-para-margin:0in;
    mso-para-margin-bottom:.0001pt;
    mso-pagination:widow-orphan;
    font-size:10.0pt;
    font-family:"Times New Roman";}
    At the current time we have one Domino v7.0.4 server (Win2k with DUC v1.2.3) in production and our clients are running Notes v7.0.4 client with DUC v.1.2.3 installed on them. This setup has been working great in our Unity v5.0(1) environment. We are looking to migrate our users to a newly installed Domino v8.5.1 (Win 2008 x32 DUC v1.2.5) server in the near future but out test users are running into a problem.
    When a user gets migrated over to our new Domino server the client can no longer access their voicemail via TUI.  During the user migration process we do change the pointer in Unity to their mail file on the new Domino server. When the user logs in they receive "Your messages are available now"
    If we migrate the user back to the old v7.0.4 server they can access their voice mail with no issues.
    Any feedback you could provide would be great. Thanks!

    Hi Kenneth,
    Could you clarify what you mean by "During the user migration process we do change the pointer in Unity to their mail file on the new Domino server"? Typically when moving a user's mail file to a new Domino server, Unity should automatically pick up those changes. Are you pehaps creating a new mail file and pointing Domino to it instead of moving it? If that is the case, the new mail file would not have been DUCS enabled, a process it must go through before Unity will be able to access it properly. The only way to force that to happen would be to reimport the user into Unity. If the mail file is properly migrated, however, the mail file should remain DUCS enabled and there would be no need for this.
    An easy way to determine whether the user is DUCS enabled is using the DUT tool. It is located on the Unity server under \Commserver\Utilities\Domino\DUT. It will show you a list of all of your users. Compare the values you see in the UCProfile section for a working / non-migrated user to a failing one. You might also want to verify that it lists the updated mail server and mail file for the user.
    Also, while I don't think it would cause this particular problem, you should be aware that Server 2008 is not currently a supported platform for DUCS/CsServer 1.2.5, see page 6 in the admin guide:
    http://www.cisco.com/en/US/docs/voice_ip_comm/unity/duc/admin/guide/cuducag.pdf
    Hope this helps,
    Pat

  • Application Server Session timed out (-11) (lack of debug in internal ITS)

    Dear Sirs,
    I get the error message below, when I try to run a IAC through internal ITS. The IAC worked for a while (<30 minutes), and then suddenly this error appeared.
    I can see that this error message has been asked about previously, but then in connection with statefull/stateless BSP applications (e.g "Session timed out thread", on this forum May 31 2005).
    In this thread the solution is apparently to alter the 'rdisp/plugin_auto_logout', is this the case for us as well?
    I have read Error handling using the ICM , but not able to actually beeing able to pinpoint the problme.
    All our other IAC are working it is only a problem with one of them. The problem is still valid even if I log on with different user, and the error is still there now even after two hours. The problem transaction is a altered version of the WSTA, and I am able to get the first screen where I enter the store number. 
    Any tips or suggestions to pinpoint the error?
    +400 Session timed out - please log in again
    Application Server Session timed out (-11)
    Error:     -11
    Version:     6040
    Component:     ICM
    Date/Time:     Thu Apr 27 14:13:56 2006
    Module:     icxxthr_mt.c
    Line:     1719
    Server:     xxXXXt_TXX_XX
    Detail:     Session does not exist+
    In SM21 :
    +Transaction Canceled ITS_P 025 ( 0x2103: Interpreter: The input contains errors. )                                                              
    Documentation for system log message D01 :
    The transaction has been terminated.  This may be caused by a termination message from the application (MESSAGE Axxx) or by an error detected by the SAP System due to which it makes no sense to proceed with the transaction.  The actual reason for the termination is indicated by the T200 message and the parameters.                         
    Additional documentation for message ITS_P               025 Template interpretation terminated - internal error &1                        
    No documentation exists for message IT025+                                     
    Best regards,
    Jørgen

    Dear Sirs,
    The error has been solved. It was a matter of an illegal input sendt to the RetailHeaderFunction.
    However, this makes way for another question.
    On the standalone ITS, it was possible to get error messages which indicated where the code contained errors, when it was runned.
    How is it possible to make the internal ITS gives better feedback (errors?), and show where the error is?
    Jørgen

  • Timeout when retrieving document from Content Server

    Dear experts,
    after the migration of an R/3 Enterprise system with a file-based Content Server (CS without database) we experience significant delays when a document shall be retrieved from the Content Server and displayed.
    The repository, from which documents are retrieved, has the following properties:
    Document Area = ArchiveLink
    Storage Type = HTTP content server
    Obviously the delay is caused by time-outs when R/3 tries to access the content sever.
    The RFC trace recorded during the document retrieval contains the following entries:
    Trace file opened at 20100825 104048 CEST SAP-REL 640,0,304 RFC-VER 3
    Trace file opened at 20100825 104115 CEST SAP-REL 640,0,304 RFC-VER 3
    End of trace *****
    ======> CPIC-CALL: 'ThSAPCMRCV' : cmRc=20 thRc=456
    Timeout beim Verbindungsaufbau (Partner vorhanden ?)
    ABAP Programm: SAPLSCMS_CLTFC (Transaction: )
    Called function module: RFC_PING
    User: TSI0403 (Client: 007)
    Destination: SAPCMS (handle: 4, , {4C74C1A6-BD3B-3B59-E100-00001A5F0A18})
    SERVER> RFC Server Session (handle: 1, 41571399, {4C74D031-BD2C-3B58-E100-00001A5F0A18})
    SERVER> Caller host: cimpd38_MPD_38
    SERVER> Caller transaction code: MIR4 (Caller Program: CL_GOS_TOOLBOX_VIEW===========CP)
    SERVER> Called function module: GOS_EXECUTE_SERVICE
    Error RFCIO_ERROR_SYSERROR in abrfcpic.c : 3268
    CPIC-CALL: 'ThSAPCMRCV' : cmRc=20 thRc=456
    Timeout beim Verbindungsaufbau (Partner vorhanden ?)
    Error RFCIO_ERROR_MESSAGE in abrfcio.c : 1660
    ======> CPIC-CALL: 'ThSAPCMRCV' : cmRc=20 thRc=456
    Timeout beim Verbindungsaufbau (Partner vorhanden ?)
    ABAP Programm: SAPLSCMS_FE_RFC (Transaction: )
    Called function module: RFC_PING
    User: TSI0403 (Client: 007)
    Destination: SAPHTTP (handle: 6, , {4C74C1B4-BD3B-3B59-E100-00001A5F0A18})
    SERVER> RFC Server Session (handle: 1, 41571399, {4C74D031-BD2C-3B58-E100-00001A5F0A18})
    SERVER> Caller host: cimpd38_MPD_38
    SERVER> Caller transaction code: MIR4 (Caller Program: CL_GOS_TOOLBOX_VIEW===========CP)
    SERVER> Called function module: GOS_EXECUTE_SERVICE
    Error RFCIO_ERROR_SYSERROR in abrfcpic.c : 3268
    CPIC-CALL: 'ThSAPCMRCV' : cmRc=20 thRc=456
    Timeout beim Verbindungsaufbau (Partner vorhanden ?)
    Error RFCIO_ERROR_MESSAGE in abrfcio.c : 1660
    I wonder why R/3 first attempts to contact the Content Server via the SAPCMS RFC destination rather than using HTTP (SAPHTTP)? - The connection tests for both RFC connections are successful.
    Is there a way to instruct R/3 to go via the SAPHTTP RFC destination as the first choice and NOT via SAPCMS when documents shall be retrieved and displayed?
    Thank you,
    Rainer Walter
    Edited by: Rainer Walter on Sep 3, 2010 3:07 PM

    Hi Ravindra,
    thank you for looking into this problem. The Content Server is not based on SAP DB/MaxDB. The repositories are represented as plain directory structures in a file system.
    My issue is that R/3 is not using the SAPHTTP RFC connection as first choice although the repository is flagged as "HTTP Content Server" but uses another RFC destination instead. When this attempt is timed out, R/3 eventually picks the appropriate RFC destination. So it takes a considerable amount of time until a requested document is finally retrieved from the repository in the Content Server.
    I am confused because the system where I experience this behaviour has been built from a system copy of a production system, and the Content Server for this new system has been built from scratch. In the original production system, no timeout is found when users are retrieving documents, and obviously the original R/3 system is using SAPHTTP right away. The original production system and the new system built from a system copy are accessing different Content Servers.
    I am looking for a way to instruct the new R/3 system to immediately use the SAPHTTP connection when talking to the Content Server, and I would be grateful for a hint where I can configure this to get rid of these timeouts.
    Best regards,
    Rainer

  • When we have to go for session method, when we have to go for call transact

    when we have to go for session method, when we have to go for call transaction method if i have a 3000 records in flat file. which is better? why

    Data Transfer
    During the process of data transfer, data is transferred into the SAP R/3 System. This transfer is from an external system to SAP R/3 system. Whenever you transfer data from an external system into an R/3 System, you can use data transfer because it is installed and regularly transfers data from an external system into an R/3 System.
    As discussed, with the help of BDC, you can transfer the required data from a non-SAP system to an SAP system. For this kind of data transfer you are required to write an ABAP program. This ABAP program would help to export the concerned data to a sequential dataset file. The data in this file has to be stored. This should be stored in a format, which is acceptable to SAP batch input program. But, to transfer data from a SAP system to another SAP system, you can take the aid of RFC or CPI-C.
    SAP application supports the data transfer of numerous SAP business objects. The said data transfer program specifies the data format definition, which is necessary to import the data into the R/3 System. There are three methods available for transferring data:
    Direct Input:
    In this method the SAP function modules execute the consistency checks. However, there are other means of checking with the help of screens. The Direct Input Method has considerable performance advantages.
    Call Transaction:
    In this method you can check the data consistency with the help of screen logic.
    Batch Input Session:
    In this method data consistency is checked with the help of screen logic.
    Direct Input Method
    Among the methods of data transfer through BDC, direct input method is the one that is used, especially in case of transferring large amount of data. In order to enhance the batch input procedure, the system offers you with the direct input technique.
    There is a distinction between the batch input technique and this technique. Unlike batch input technique, this technique does not create sessions. Instead, it stores the data directly. Moreover, it does not process screens. The data has to be entered directly into the corresponding database tables. The system calls a number of function modules which execute necessary checks, if any required. In the case of errors, the direct input technique has a facility to restart the entire mechanism. However, if you want to restart the entire mechanism in case you faced an error, then direct input programs must be executed in the background only. One has to use program RBMVSHOW or Transaction BMV0 to maintain and start these programs.
    Call Transaction Method
    Call Transaction method is another method used for Data Transfer. In this type of method your program will use the ABAP statement CALL TRANSACTION USING in order to run a SAP transaction. In this type external data need not be deposited in a session for being processed later on. Instead, the entire batch input process takes place inline in your program.
    Here, the data transfer program must convert the data that has to be transferred into the SAP system. This is as per requirement by the SAP data structure or the transaction which is using it. It is to be remembered that a conversion of the data types may be necessary at different times during the process.
    Suppose there is a data type mismatch then you have to convert the data types to type C. In this regard the data transfer program should be capable of exporting the data in SAP format to the sequential file. At the time of uploading the data into the SAP system, the BDC program reads the data from the abovementioned sequential file.
    Batch Input Session Method
    This is the third method for data transfer. If you use the batch input method to transfer data, then you should remember that an ABAP program has to read the external data which is to be entered in the R/3 System. Subsequently, it stores the concerned data a "batch input session." The batch Input session records the actions which are required in the process of transferring data into the system. This can be done by using normal SAP transactions.
    As soon as the program generates the said session, you will be able to run the session in order to execute the SAP transactions in it. Moreover, you can start the session, and at the same time, can monitor a session with the help of batch input management function. For this you have to choose:
    System à Services à Batch input. Moreover, you can have the session run in the background processing.
    Writing a Data Transfer Program
    If you want to write a data transfer program, you have to follow the steps mentioned below.
    Firstly, you will analyze the structure of the existing data. Subsequently, your job is to specify the conversions, which are essential to fill the SAP data structures.
    Secondly, you have to generate the SAP data structure. In case the program is written in ABAP, you will require only the required tables in the concerned program with the help of TABLES statement.
    Thirdly, you will have to initialize the SAP data structure.
    Fourthly, fill the structure with data, performing any conversions and error checking that are required.
    Finally, you will write the sequential file. In the SAP system this sequential file is typically required for making the data available to the batch input program.
    Batch Input Method
    Batch input method is a type of data transfer method. It is used for bulk data transfer; it is one of the primary ways by which data can transferred into the R/3 System. This method is not for near real-time data transfers.
    There are various typical uses of batch input. One of the ways includes the one-time import of data. This import of data is from a legacy system into a newly installed R/3 System. In addition to it, another typical use is for periodic (i.e. hourly, daily..., and so on) transfers of data. These transfers are from external systems or legacy systems which are still in use into R/3 system where all enterprise data is consolidated.
    The R/3 applications deliver different programs for batch input, which are ready to be used. However, in some cases a customer has to write his or her own batch input program. This is required in order to convert the concerned data from a legacy System or from a proprietary format into an R/3 data format.
    The process flows for a batch input are discussed below.
    Data Transfer Decision-Making: It is with a decision to transfer data from an external source into R/3 that the process of batch input begins. It is probable that the external source may be a legacy system that is being replaced. A one-time bulk data transfer is foreseen in this regard. Alternatively, the external source may be an external system that is to remain in use. In this case, a regularly recurring bulk data transfer is foreseen.
    Setting up Batch-Input for Data Transfers: If R/3 standard one-time or regular data transfers are required, then by means of customizing settings in the R/3 Customizing System in SAP ASAP set up will occur. You must set up custom batch input procedures by hand, which means the system administrator must schedule the data conversion program that creates the batch input session. The system administrator and the batch input programmer must determine the following: how frequently data is made available from the external system, how frequently the conversion program should run, and whether the conversion program runs in R/3 (ABAP program) or in a host system (external program).
    Processing Batch Input Sessions: When a batch input session is processed, then the actual transfer of data into R/3 takes place. Little attention is required in processing of batch input sessions by the system administrator. Usually, the starting of batch input sessions is automated by the system administrator. If necessary, the administrators can also start batch input session explicitly from transaction SM35.
    Checking Batch Input Sessions: For a system administrator the routine activity is to check daily or more frequently in transaction SM35 whether all batch input sessions have been completed successfully. It is the schedule for running batch input sessions on which the schedule for checking sessions depends upon. For doing this check the R/3 System provides easy-to-use batch input management tools.
    Analyzing Errors: It is the duty of the system administrator to analyze the problem if one or more transactions in a session end in errors. Usually, the assistance of the affected data entry specialist or department for this analysis will be needed by the system administrator. In the situation where the problem was caused by incorrect data conversion or incorrect generation of the batch input session then the programmer who wrote the data conversion program may also need to be involved.
    Error Handling in Batch Input Method
    It is found that most problems usually fall into one of the following two categories discussed below.
    In this case either required data is missing from the batch-input session or invalid data has been included in the session. Errors in the data conversion program or the presence of unexpected types of data or incorrect data in the legacy database are the possible external causes of this type of problem. Within R/3, the causes for this type of problem include incorrect or incomplete customizing in an application. For example, a legacy data type may not have been foreseen in the check table entries made in application customizing.
    This case mainly includes technical/programming problems. The data is entered by a batch input session by running R/3 transactions non-interactively. Therefore, a typical technical or programming problem is the incorrect identification of one of the data fields in a transaction. Thus, the conversion program may not fill a required data field or may have provided invalid values.
    Conclusion
    I have gone through the details of batch data communication and various methods used to transfer data. One can make use of BDC to transfer data from a SAP to SAP system. In addition, it can transfer data from a non SAP system to SAP system too. I have discussed the various methods of BDC and the error handling in the respective methods.
    Thanks,
    Shankar

  • When we go for shared Server environment?

    When we go for shared Server environment from Dedicated server mode?
    When we have to choose thi shared Server mode?
    Which is the optimized way?

    > When we go for shared Server environment from Dedicated server mode?
    You do not. It is not one or the other - it is using both correctly. You want shared server sessions to service applications (any, from web/app servers to flat clients) that are OLTP in nature. They fire off short-and-sweet SQLs that takes a second or three to execute.
    A shared server pool of a few processes can service a very large number of clients.
    You want to use dedicated servers to service OLAP type clients. They fire off complex SQLs that can tie up the server processes for several minutes at a time.
    And that in a nutshell is how to apply the two. It is not Shared Servers versus Dedicated Server. It is about using The Right Tool For The Job. Nothing more and nothing less.
    As for Windows, I will not really bother quickly with shared server connections. On Windows both shared and dedicated servers runs as threads in the main Oracle executable process image.
    There is thus very little resource saving by reducing the number of threads as the footprint of a thread is tiny in comparison to a brand new process. The amount of resources saved by a reducing the number of dedicated server threads is not significant. If anything, Oracle on Windows has shown to scale better than Oracle on Linux in this regard. (refer to [url http://www.perftuning.com/pdf/Comparison_Oracle_Windows_Linux.pdf]A Comparison of Oracle Database 10gR2 Real Application Cluster Performance on Microsoft Windows 2003 Server Enterprise Edition x64 and Red Hat Enterprise Linux x86_64 from the Performance Tuning Corporation).
    And this is exactly the opposite of how Oracle behaves on Unix. Each server process is a physical kernel process. This means that a servicing a 1000 concurrent sessions with a 1000 dedicated server processes, each consuming memory and resources, are expensive. Conversely, supporting a 1000 concurrent sessions with a 100 shared server processes is (almost) 90% cheaper.
    This makes Shared Server a desirable option to use to scale OLTP sessions on Unix-based systems (or any platform where Oracle does not use multi-threading but multi-processing).
    PS. As the Linux kernel support Posix threads (see http://en.wikipedia.org/wiki/Native_POSIX_Thread_Library), it begs the question as to when Oracle is going to support multi-threading on Linux instead of multi-processing.. or at least give us the option to choose.

  • Terminal Server Session Freezes/Hangs

    Hi,
    Just a little background info, I have a terminal server farm running Windows 2003 R2 Enterprise behind a hardware load balancer and I am utilizing the session directory service.
    The issue I am running into is users’ terminal sessions will freeze up when they reconnect to their existing session. There are two scenarios that this happens. When a user moves from one desk to another and they log into the terminal server and reconnect
    to their session or after a period of inactivity their session will lock them out and then they log in. All the user gets is usually only a partial display or just the desktop backgroup and the session doesn't respond. I used to just kill their session and
    have them log back in but now I have narrowed it down to the OUTLOOK.EXE process. If I log into the server they are on and kill their Outlook process their session starts to respond. I have noticed that Outlook will be utilizing 2% to 10% CPU when it is occurring.
    We are running Office 2010. It seems to be random and very frustrating.
    Any help would be greatly appreciated.
    Thanks

    Hi,
    Thanks for your posting.
    According to your issue, I can let you know some solutions but before that want information from your side.
    a. How many clients are logged in TS session at a time?
    b. Have you tried by disabling Offloads?
       Information regarding Offloads:
       http://support.microsoft.com/default.aspx/kb/951037
    c. You check the performance of your server by using Performance Monitor tools and can capture logs.
       http://technet.microsoft.com/en-us/library/cc749115.aspx
    Please install the required hotfixes available for the issue.
    1. Terminal Server and connected Terminal Services clients pause when a Terminal Services client logs on or logs off (KB 324446):
       http://support.microsoft.com/kb/324446
    2. A Windows Server 2003-based or Windows Server 2008-based terminal server stops accepting new connections, and existing connections stop responding (KB 956438):
       http://support.microsoft.com/kb/956438
    Might This Helps!
    Thanks.

  • Firefox keeps causing internet conenction to drop when I try to load a session

    I'm using Firefox 11 on Windows Vista Home Premium 64-bit, and I'm using a cable modem and a wireless router, with my computer connected to the router via Ethernet, and this problem is driving me absolutely INSANE! When I try to reload a session in Firefox, it causes my Internet connection to drop, and I end up having to power cycle my router, and often my modem, too. What in the world could cause it to this? I've tried different modems. I've tried different routers. I've tried reducing the number of tabs in the session. No luck. I've already wasted an awful lot of time on this, and I'm looking for some targeted, specific assistance regarding what might be causing this to happen, and what I can do about it. I don't want to waste too much time shooting around in the dark, just trying enabling and disabling different extensions at random, for instance, to see if that might be causing the problem. I am EXTREMELY frustrated and upset, and I'm in no mood to go through the aggravation and wasted time of all that blind flailing. So, what specific things could be causing my problem, what specific things should I be looking for, and what specific things should I try in order to attempt to remedy this situation? For instance, might it be something like limiting the number of pages it loads simultaneously while it's loading a session, and, if so, how do I do that? Or am I completely off base on that speculation? Again, I'm looking for some targeted assistance in order to try to limit the amount of wasted time and aggravation I have to put up with. So, can anyone help me with that?

    Something you installed has changed your user-agent. In your post under "User Agent", the "YFF3" is suspicious.
    See: http://en.wikipedia.org/wiki/User_Agent
    <u>'''Reset general user-agent'''</u>
    -type '''about:config''' in the URL/address bar and press the Enter key
    -if you see the warning, accept it (promise to be careful)
    -Filter = general.useragent.extra.firefox
    -the item will be '''bold''' and show '''user set'''
    -right-click that preference and select Reset
    -restart Firefox (File > Restart Firefox)

  • ORACLE server session terminated by fatal error

    Hi,
    When i try to startup my database(10g), i am getting this error after database monted.
    ORA-00603: ORACLE server session terminated by fatal error.
    (Startup nomount is ok)
    conn /as sysdba
    Total System Global Area 1048576000 bytes
    Fixed Size 1223392 bytes
    Variable Size 343934240 bytes
    Database Buffers 700448768 bytes
    Redo Buffers 2969600 bytes
    Database mounted.
    ORA-00603: ORACLE server session terminated by fatal error
    Alert.log:
    Errors in file /u01/app/oracle/product/10.2.0/db_1/admin/db/udump/db_ora_9917.trc:
    ORA-00603: ORACLE server session terminated by fatal error
    ORA-00600: internal error code, arguments: [4194], [24], [14], [], [], [], [], []
    ORA-00600: internal error code, arguments: [4194], [24], [14], [], [], [], [], []
    The trace file is 1,6GB(!)
    Trace file:
    ----- Redo read statistics for thread 1 -----
    Read rate (ASYNC): 26Kb in 0.78s => 0.03 Mb/sec
    Total physical reads: 4096Kb
    Longest record: 0Kb, moves: 0/62 (0%)
    Change moves: 35/122 (28%), moved: 0Mb
    Longest LWN: 14Kb, moves: 0/13 (0%), moved: 0Mb
    Last redo scn: 0x0000.31539752 (827561810)
    NO VALID LOG MEMBER FOR SEQ# 20884 OF THREAD 1!
    NO VALID LOG MEMBER FOR SEQ# 20883 OF THREAD 1!
    *** 2011-01-06 06:18:51.782
    ksedmp: internal or fatal error
    ORA-00600: internal error code, arguments: [4194], [56], [49], [], [], [], [], []
    Current SQL statement for this session:
    update service$ set name = :1, name_hash = :2, network_name = :3, creation_date = :4, creation_date_hash = :5, deletion_date = null, goal = :6, flags = :
    ----- Call Stack Trace -----
    Thanks

    Hello,
    ORA-01194: file 1 needs more recovery to be consistent
    ORA-01110: data file 1: '/u01/app/oracle/product/10.2.0/oradata/db/system01.dbf'The system01.dbf which belongs to the Tablespace SYSTEM is still inconsistent.
    So, I think you'll need to start from a Backup and apply all the Archived Redo logs before the ORA-600 occured.
    More over, try to open a Service Request to My Oracle Support and, send them all the Traces Files and Alert Log so that they can follow all the actions which has occured on this Database. You may need their help.
    Best regards,
    Jean-Valentin

  • When should use request scope

    hi
    i don't know when i should use request scope, the example is use session scope always.
    Can anyone help me?
    thanks!

    SpiderFromMars wrote:
    > What are the best practices for using the variables
    scope versus using the
    > request scope when assigning variables.
    >
    > In other words, when should I use <cfset variables.x
    = 'barf'> and when
    > should I use <cfset request.x = 'barf'>?
    >
    The variables scope is only accessible in the current
    template. The
    request scope is accessible in the current template and any
    other
    template, such as a custom tag or include file, that is used
    to build
    the current request.
    I.E
    <!---index.cfm--->
    <cfset variable.x = "foo">
    <cfset request.x = "bar">
    <cfoutput>#variable.x##request.x#</cfoutput>
    <cf_myCustomTag>
    <!---myCustomTag.cfm--->
    <cfset variable.x = "bob">
    <cfoutput>#variable.x##request.x#</cfoutput>

Maybe you are looking for

  • I can not delete bookmarks from the bar as I am spammed with many bookmarks.

    On my bookmark bar I have multiple instances of "Webmail Notifier" and they are in my bookmark list in many places. I am unable to delete them at all. How can I fix this issue? I have many bookmarked items that I do not want to lose as well.

  • Business Rules in Planning 11.1.2.1

    Hi, I'm having a strange problem in Planning 11.1.2.1 dataforms. I added a business rule and selected Run on Save, Use members in dataform and Hide Prompt in a dataform. However when I press the save button, the Business Rule prompt appears, and the

  • PC Will Not Boot With Sound Blaster Live Va

    My PC won't boot with the SBLV card but when i put the SB Vibra 28 card back in it will boot. The card dosen't look burnt out anywhere,the sound card could have **bleep** itself,any ideas.also i have another problem i can't get the SB Vibra 28 driver

  • Flash based uploader script problem. V18.0 came with this bug!

    i am a web developer. everything works fine on firefox every time and my first browser which i always use it is firefox... but only a problem, never solved. FLASH UPLOAD PROBLEM.. there is a IO Error problem and it has never been solved... But with 1

  • Question on CCM catalog content approval

    Hi, Can somebody tell me what kind of update to catalog content needs approval? My understanding was that if there is a price change, definately it will need to be approved. But yesterday my catalog manager uploaded some change to his catalog item pr