Dynamic Memory in the production environment

Hi,
We configure all production VMs with static memory  on Hyper-v 2012
we need to make sure there is no negative impact on VM performance if we use Dynamic memory , as we read for 2 years
not recommend to use dynamic memory in the production.
I preferred the answer with Trusted URL
Ramy

Hiya,
The best answer you can get is, that it is dependent on the application running on the server O/S.
The reason that dynamic memory is not recommended for production environments, is that many applications are not supporting it. That is usually seen on memory intensive applications(SQL is an example) or simply because of caching types of functions. (SharePoint
is an example).
The major concern here is when VM's are decreasing the memory for a VM, the application does not understand this. Usually it is not a problem when increasing the memory. - Again it will depend on the application.
Think of it as hot swap memory and how you used to use that in the physical machine days.
Also when there is not adequate memory, operating system will use paging. Paging uses disks and disks has a lower access time than memory.
In general dynamic memory is easier to control than dynamically expanding disks, as you can set the buffer size of the memory, which is still not available for the dynamic disks.
Besides the above, the following links states:
"Workloads that are not NUMA-aware will not take advantage of virtual NUMA. However, the guest operating system may perform some NUMA optimization. Enabling Dynamic Memory (therefore presenting only a single virtual NUMA) should not cause performance
degradation"
http://technet.microsoft.com/en-us/library/dn282282.aspx

Similar Messages

  • What authorizations to give to developers in the production environment

    Hi,
    We have setup our development PI 7.1 environment and are now ready to setup production.  Can someone point me in the right direction for determining which authorizations need to be given to the PI developers in the production environment?
    Do they have to be the same or is there some other set of roles/authorizations for a production environment?
    Any help is greatly appreciated.
    Best Regards,
    Rick

    Hi Rick,
    Precisely for configuration XI have separate role: SAP_XI_CONFIGURATOR which is used to configure interfaces in ID. But again this is role provided to very selective people (usually Application Support team) because you need to share some sensitive user/password etc for different systems (SAP/Non-SAP).
    We used following roles combinations:
    - Dev system
    Developers = SAP_XI_DEVELOPER, SAP_XI_CONFIGURATOR, SAP_XI_MONITOR, SAP_XI_BPE* ,
    TechLead = Developer + SAP_XI_ADMINISTRATOR
    - QA System
    Developer = SAP_XI_CONFIGURATOR, SAP_XI_MONITOR, SAP_XI_BPE_MONITOR
    TechLead = Developer (in special cases SAP_XI_ADMINISTRATOR)
    - PROD System
    Developer = SAP_XI_MONITOR, SAP_XI_BPE_MONITOR, (None  for Configuration If I have separate team for configuration).
    TechLead = Developer +  SAP_XI_CONFIGURATOR
    It is best practise to keep separate team for configuration on Prod system because of sensitivity of information. Even on PROD system you need to modify SAP_XI_MONITOR role to disallow monitoring of payload because of data sensitivity (HR, Finance data etc).
    Regards,
    Gourav

  • Use of Emigall for creation of masters in the production environment

    Hi,
    The EMIGALL objects are normally used for migrating legacy master data and/or cut-over data before production environment is up.
    I am contemplating to use EMIGALL object for creation of contract account masters in production environment. One more option that I have is to use standard BAPI for creating contract account masters.
    Can anyone tell me whether it is proper to use EMIGALL object for day to day creation of master day in production environment. Is there any disadvantage or risk involved in it.
    Kindly reply soon.
    Regards,
    Ganesh

    I've already used emigall multiple times to do delta migrations into an operational prod environment.
    Purely looking at the functionalities, it should/would/must be possible to use emigall as a master data generator. I just think you need to look into the requirements:
    who will use it? end-user/application manager/...
    what's the amount of data to be loaded?
    what's the time window of the load? Day/night
    how is the data supplied?
    As you know, emigall EATS system resources like mad, so using it during the day might not be preferable. emigall is also very picky about the file format, whereas in a custom report you can define the input structure yourself.
    On the other hand, the error handling and follow-up of emigall is great...
    Personally, I'd go for a custom report with a BAPI... It'll give you more flexibility than emigall.

  • Unable to import the Internet servivce into Production Environment

    Hi,
    We put the internet service and its HTML templates in a custom package for integrated ITSMobile. We got everything but the internet service ZXXX in the production. Has anyone come across this before? We were able to transport to the quality system. But it did not create the internet service in the production environment. Is it authorization issue? Your input will be highly appreciated.
    Regards,
    Subbu

    Hi Subbu,
    I suggest you to as your basis guys how are familiar with transports between systems. They will find out what the reason is. Without being able to look into transport protocols it is unlikeliy that someone here can advise you.
    Regards,
    Klaus

  • Dynamic Memory and Data Deduplication

    Hi,
    I'm currently testing W2K12 Deduplication feature on some servers, and I was wondering if I could enable "Dynamic Memory" on the VM in Hyper-V? Because the job only uses percentage of memory for the dedup job, I guess it's not possible.
    Thanks in advance!

    I know it old thread, but I ran into this and could not find anything on it but here, so I am adding my experience.
    I can concur with 3magroup...
    In my environment Windows Server 2012 with Dynamic Memory and Deduplication enabled, the HyperV does not release the driver locked ram for the dedup optimization. What resulted was an insane amount of paging and a painfully slow VM. Increasing the memory
    buffer for the dynamic memory did not cause the HyperV to release the memory for the dedup. Instead, it still "driver locked" nearly all of the ram, which did not make sense to me based on my understanding of the Memory Buffer setting.
    I can verify the VM poor performance and dedup using the paging file with the Resource Monitor, disk section, and the Driver locked RAM with RamMap.

  • Exporting snapshotted VMs and dynamic memory

    We have a production VM which uses dynamic memory. Unfortunately, someone decided to snapshot this machine and now we cannot export the VM using the hyper-v "Export" facility.
    Export gives a warning message indicating that VMs and snapshots cannot be imported into Server 2008R2 and that dynamic memory must be turned off prior to exporting. Here's the message:
    I can turn off dynamic memory in the main VM but each snapshot has its own setting for dynamic memory. I can see the setting, dynamic memory is on but it is grayed out.
    As a backup, I want to export the VM prior to merging the snapshots. How can I do this?
    Thanks,
    Alan
    [email protected]
    Alan MacKenzie [email protected]

    Yes, you cannot import to Server 2008 R2, you can only import to Server 2008 R2 SP1 or newer.
    You cannot go backward in the version of the hypervisor, that is what it is warning about.  It is not a block, it is just information.  You can still Export and Import (if necessary) to a 2008 R2 SP1 or 2012 machine.
    In regards to the snapshot.  That is a different issue.  Simply delete the saved state portion (the running memory part) of the snapshot.
    Brian Ehlert
    http://ITProctology.blogspot.com
    Learn. Apply. Repeat.
    Disclaimer: Attempting change is of your own free will.

  • Production environment, please help...

    Hi,
    we have a large BC4J web based application.
    we plan to put it production to allow to 40 CONCURENT USERS to connect to...
    after testing the application in stand-alone OC4J. it works well, nut at some moments, the server print the message
    like out of memory and it becomes realy slow. i restarted the OC4J
    like java -Doc4j.userThreads=true -Xms100m -Xmx800m oc4j.jar to increase the heap and i hope that this will
    solve the problem.
    no for the production environment :
    hardware
    Sun server
    4 x CPU x 450
    4 G on RAM.
    software
    Sun Solaris 8
    Oracle 9iAS R 2 (9.0.2.0.0)
    so what is the good configuration for the OC4J in 9iAS to get our system support all users. should we
    use - Doc4j.userThreads=true -Xms100m -Xmx800m even whith 9iAS.
    thank you for your help.

    El Bachir,
    The heap size is dependent upon your application and no. of concurrent users you are using. Yes, you can specify the the heap size for OC4J also in an Oracle9iAS instance from the EMWebsite by navigating OC4J Instance ->Server Properties -> specify the parameters for the Java command.
    For details how to determine the heapsize please look at the Performancing Guide. Here is the link
    http://otn.oracle.com/docs/products/ias/doc_library/90200sol_otn/perf.902/a95102/optj2ee.htm#1012258
    regards
    Debu Panda
    Oracle
    Hi,
    we have a large BC4J web based application.
    we plan to put it production to allow to 40 CONCURENT USERS to connect to...
    after testing the application in stand-alone OC4J. it works well, nut at some moments, the server print the message
    like out of memory and it becomes realy slow. i restarted the OC4J
    like java -Doc4j.userThreads=true -Xms100m -Xmx800m oc4j.jar to increase the heap and i hope that this will
    solve the problem.
    no for the production environment :
    hardware
    Sun server
    4 x CPU x 450
    4 G on RAM.
    software
    Sun Solaris 8
    Oracle 9iAS R 2 (9.0.2.0.0)
    so what is the good configuration for the OC4J in 9iAS to get our system support all users. should we
    use - Doc4j.userThreads=true -Xms100m -Xmx800m even whith 9iAS.
    thank you for your help.

  • Issue in production environment

    Hi,
    We got an issue in the production environment on last Friday and yet not able to find out the root cause of the issue.
    Environment
    GUI : .Net
    Server : Java
    Application Server: Jboss
    Database: Sqlserver
    The Java application is deployed into Jboss application server which connects to SQL server as back end. GUI makes the web service call to server for data communication
    Issue
    Friday at 2 PM users reported the slowness of the application (Not getting the response from server in GUI), all the request from GUI were getting timed out.
    Restart of Jboss didn't help
    Restarted the Jboss and sqlserver for the second time and than the environment became stable
    Analysis
    1. From the thread dump of Jboss log we see that there are many threads waiting on a socket for database connection(according to database team all the connections are open and available at that particular time).
    2. The size of transaction log is almost doubled during this period (when the issue was reported). 
    We couldn't find a reason why this issue is happened. Is it a database issue or something else? Please suggest...
    Thanks,
    Manoj

    Hi Manoj,
    According to your description, when running the web application, many threads are waiting for the socket from database which cause requests from front GUI time out. Right?
    In this scenario, user can access the front GUI, which means the application server is working. Since the connections are all open at that particular time, it should not be issue on your JDBC. It seems to be a deadlock issue which cause other threads
    waiting and hang. I recommend check and optimize your code. On database side, I suggest you open the SQL profiler, select DeadLock Graph, it will record when deadlock occurs. Please refer to links below:
    Detecting and Ending Deadlocks
    Analyze Deadlocks with SQL Server Profiler
    If you have any question, please feel free to ask.
    Simon Hou
    TechNet Community Support

  • How to port database changes from development to a production environment

    How do I port database changes from the development to the production environment?
    I am using v8 and have always had to redo everything using the schema manager all over in the production environment. Is there an easy way to generate a script, for example to dump the database changes on the development machine to be executed later on the production machine?

    This should already be a clearly defined change control process. Once a procedure, function, package, trigger, or whatever completes the testing rounds, it should be promoted to production.
    Forgive me if it seems I'm trivializing, but I don't see the problem, just copy the object(s) from your software library (or development) into production using whatever tool works best or has been chosen. If you are doing data copies then you have various options again including good old export/import.

  • Production Environment RFC's for EWA.... Please help!!

    We have some of the production environment setup EWA. I will need to add else which are left. I have setup EWA in sandbox for testing. worked fine. but i have question about RFC's.
    My all DEV systems has the following RFC's in solman;
    SM_XXXCLNT100_LOGIN
    SM_XXXCLNT100_READ
    SM_XXXCLNT100_TMW
    SM_XXXCLNT100_TRUSTED
    But my all QAS and PRODUCTION system has the following RFC'c in solman.
    SM_XXXCLNT100_LOGIN
    SM_XXXCLNT100_READ
    SM_XXXCLNT100_TMW
    I thought when I generate auto RFC's from SMSY, it creates the following;
    SM_XXXCLNT100_READ
    SM_XXXCLNT100_TMW
    SM_XXXCLNT100_TRUSTED
    What should I do? if it creates _TRUSTED ONE in PRD, should I delete it? Please help!!
    Question: If I am generating auto RFC's for PRD, it created _LOGIN one too? I am totaly confused??
    Thanks,
    I will definitely post points

    The only reason I want to do that because I have to talk to security folks for that, they will go and look at the other production environments which dont have _TRUSTED rfc...
    Like as I have mentioned, all production has the following rfc's
    SM_XXXCLNT100_LOGIN
    SM_XXXCLNT100_READ
    SM_XXXCLNT100_TMW.
    I dont understand only one thing. how can I get _LOGIN rfc?
    The only reason I am asking, when I did EWA in sandbox, It did not generate _LOGIN rfc...
    Do you have any idea?
    Thanks,

  • Weblogic 8.1 Server log size increase in Production environment

    Hi,
    Issue:: One of the log file is increasing in size and exceeding beyond the size mentioned in the configuration file resulting in application outage.
    Issue description:
    We are having problems with the log size in the Weblogic 8.1 server. The fileminsize has been mentioned in the config.xml.
    New log files like MYsvr.log00001,MYsvr.log00002, MYsvr.log00003, MYsvr.log00004 etc are also being generated appropriately when the max file size has been reached. But simultaneously, one of the files is growing in size, exceeding the limit mentioned in the configuration file. Eg.. the MYsvr.log00001 file is 800MB in size while the other files(MYsvr.log00002, MYsvr.log00003 etc are 10MB in size)
    This increase in size of the log has been resulting in an application outage.
    More Details:
    1. Server: BEA Weblogic 8.1 server
    2. Log size is fine in other environements. This is a problem only in the production environment.
    3. The entry in the config.xml is as follows:
    <Server ListenPort="6313" Name="MYsvr" NativeIOEnabled="true" TransactionLogFilePrefix="./">
    <ServerStart Name="MYsvr"/>
    <Log FileMinSize="10000" FileName="MYsvr.log" Name="MYsvr"
    NumberOfFilesLimited="true" RotationType="bySize"/>
    <SSL Name="MYsvr"/>
    <ServerDebug Name="MYsvr"/>
    <WebServer Name="MYsvr"/>
    <ExecuteQueue Name="default" ThreadCount="15"/>
    <KernelDebug Name="MYsvr"/>
    </Server>
    Could you please help with this issue ?
    Thank you.

    Can someone please provide a solution for the issue

  • RE: Production Environment Definition

    Brad,
    We use connected environments so that we do not have a single point of
    failure.
    We use multiple environments and connect them together in a star topology
    for reliability of service. Our servers (23 in total) sit out at branches
    in the back of beyond and the WAN connections between the servers are
    unreliable. One needs a reliable connection to the Name Service which sits
    on each Environment Manager. We have thus created 23 connected
    environments with an Environment Manager on each LAN. Connected
    environments are still a bit buggy but Tech Support is currently working on
    fixing the last of the problems. We are still on ver 2H15 for this reason.
    Disadvantages of this topology are that making distributions take a long
    time because referenced partitioning cannot be scripted in fscript and
    econsole only connects to one environment at a time.
    There is a Fort&eacute; consultant in Denver called Pieter Pretorius who has had a
    lot of experience with our connected environments. It may be worth
    chatting to him.
    Regards,
    Richard Stobart
    Technical Consultant for Fort&eacute;
    E-mail [email protected]
    Quick-mail: [email protected]
    Voice: (+ 27 83) 269 1942
    (+27 11) 456 2238
    Fax: (+ 27 83) 8269 1942
    -----Original Message-----
    From: Brad Wells [SMTP:[email protected]]
    Sent: Tuesday, February 10, 1998 11:52 PM
    To: 'Forte Users - Sage'
    Subject: Production Environment Definition
    Hello again,
    We are just starting to look at what it will take to setup a production
    Forte environment. I have some general questions regarding
    considerations that may affect the environment definition and thought
    maybe some of the more experienced users could share some thoughts on the
    following:
    1) What factors lead to the creation of multiple production environments?
    a. How many environments should you use in a production situation?
    b. Do people create separate environments for separate business units?
    c. Are there performance improvements to be had by restricting the
    number of server and client nodes included in a single environment?
    d. How do the performance benefits of multiple environments compare to
    the additional complexity of managing and maintaining multiple connected
    environments?
    The initial need is for an environment that will service approximately 50
    clients and contain a couple of server nodes (database and service
    related). However, as the environment grows, it could easily grow to a
    size of 600 clients encompassing approximately 15-20 server nodes.
    At this point in time, there is no need for the failover support of
    connected environments, but this is something we will need to add as the
    environment absorbs applications with high reliability needs. Should the
    environments be setup and connected right away or can this be easily
    added on an "as needed" basis? What other recommendations would you
    make?
    Has anyone taken advantage of Forte consulting services in defining the
    production environment? Where you satisfied with the results of the
    service?
    Thanks.
    Bradley Wells
    [email protected]
    Strong Capital Management, Inc
    http://www.strong-funds.com

    On Tue, 10 Feb 98 13:52:00 PST Brad Wells <[email protected]>
    writes:
    At this point in time, there is no need for the failover support of
    connected environments, but this is something we will need to add as
    the
    environment absorbs applications with high reliability needs. Should
    the
    environments be setup and connected right away or can this be easily
    added on an "as needed" basis? What other recommendations would you
    make?
    From the Forte Systems Management point of view, you can add them "asneeded"
    fairly easily.
    Now from the application source code point of view, implementing
    Fail/Over support
    is a different story... You will need to check your SO's dialog
    durations, handle
    DistributedAccessExceptions, "warm-up" your distributed references for
    F/O,
    design a solution for restoring global transient data, do lots of
    testings etc...
    So implementing Fail/Over is not only related to systems-management
    issues, it can
    have some influence on your application(s) source code.
    Hope this helps,
    Vincent Figari
    You don't need to buy Internet access to use free Internet e-mail.
    Get completely free e-mail from Juno at http://www.juno.com
    Or call Juno at (800) 654-JUNO [654-5866]

  • Relationship between Dynamic Memory Heap and Heap Data Structure

    This question is not strictly related to Java, but rather to programming in general, and I tend to get better answers from this community than any where else.
    Somehow, my school and industry experience have somehow not given me the opportunity to explore and understand heaps (the data structure), so I'm investigating them now, and in particular, I've been looking at applications. I know they can be used for priority queues, heap sorts, and shortest path searches. However, I would have thought that, obviously, there must be some sort of relationship between the heap data structure, and the dynamic memory heap. Otherwise, I can think of no good reason why the dynamic memory heap would be named "heap". Surprisingly, after searching the web for 90 minutes or so, I've seen vague references, but nothing conclusive (trouble seems to be that it's hard to get Google to understand that I'm using the word "heap" in two different contexts, and similarly, it would not likely understand that web authors would use the word in two different contexts).
    The Java Virtual Machine Spec is silent on the subject, as "The Java virtual machine assumes no particular type of automatic storage management system, and the storage management technique may be chosen according to the implementor's system requirements."
    I've seen things like:
    [of dynamic memory] "All the blocks of a particular size are kept in a sorted linked list or tree (I extrapolate that sorted tree could imply heap)"
    [of dynamic memory] "The free and reserved areas of memory are maintained in a data structure similar to binary trees called a heap"
    [of dynamic memory] "This is not related to the heap data structure"
    [of dynamic memory] "Not to be confused with the data structure known as a "heap"
    [of data structure] "Not to be confused with the dynamic memory pool, often known as TheHeap"
    At this point, I've come to surmise that some (but not all) memory management algorithms use heaps to track which (pages? blocks? bytes?) of memory are used, and which are not. However, the point of a heap is to store data so that the max (or min) key is at the root of the heap. But we might want to allocate memory of different sizes at different times, so it wouldn't make sense to key on the amount of available memory in a particular region of the free store.
    I must assume then that there would be a different heap maintained for each size of memory block that can be allocated, and the key must have something to do with the attractiveness of the particular memory block in the heap (perhaps the lowest address, resulting, hopefully, in growing the free store space less often, leaving more space for the stack to grow, or perhaps keyed based on the fragmentation, to hopefully result in less fragmentation, and therefore more efficient use of the memory space, or perhaps based on page boundaries, keeping as much data in the same page as possible, etc).
    So at this point, I have a few questions I've been unable to resolve completely:
    1. Am I correct that the heap was so named because (perhaps at one point in time), a heap is/was commonly used to track the available memory in the free store?
    2. If so, would it be correct that there would be a heap per standard block size?
    3. Also, at what level of granularity would a heap typically be used (memory page, memory blocks, individual words (4-bytes))?
    4. What would be the most likely property one would use as a key. That is, what makes the root item on the heap ideal?
    5. Would a industrial strength system like the jvm use a (perhaps modified or tuned) heap for this sort of task, or would this typically be too naive for an real world solution today?
    Any insight would be awesome!
    Thanks,
    A.

    jschell wrote:
    I think you are not only mixing terms but domains.
    For starters the OS allocs memory. Applications, regardless of language, request memory from the OS and use it in various ways.
    There are many variations of the term "heap" like the following.
    [http://en.wikipedia.org/wiki/Heap_(data_structure)]
    [http://en.wikipedia.org/wiki/Dynamic_memory_allocation]
    A java VM will request memory from the OS (from a 'heap') and use it in its application 'heap' (C/C++) and then create the Java 'heap'. There can be variations of that along the way that can and likely will include variations of how each heap is used, potentially code that creates its own heap, and potentially other allocators which use something which is not a heap.This last part, I find a bit confusing. By "use something which is not a heap", do you mean the heap data structure, or the dynamic memory pool meaning of heap? If the former, then you would be implying that it would be common for a heap data structure to be used to manage the heap dynamic memory pool. If the latter, what would this "something which is not a heap" be? The best definition of "heap" I've found simply states that it is a pool of memory that can be dynamically allocated. If there is some other way of allocating dynamic memory, then it would suggest that the previous definition of "heap" is incomplete.
    >
    So to terms.
    1. Am I correct that the heap was so named because (perhaps at one point in time), a heap is/was commonly used to track the available memory in the free store?Which 'heap'? The VM one? It is probably named that because the implementors of the Sun VM were familar with how C++ and Smalltalk allocated memory.Okay, but that begs the question, was the heap in C++ and/or Smalltalk so named for the above queried reason?
    >
    2. If so, would it be correct that there would be a heap per standard block size?Not sure what you are referring to but probably a detail of the implementation. And since there are different levels the question doesn't mean much.
    However OS allocations are always by block if that helps. After that it requires making the question much, much more specific.
    3. Also, at what level of granularity would a heap typically be used (memory page, memory blocks, individual words (4-bytes))?Again not specific enough. A typical standard implementation of heap could not be at the word level. And it is unlikely, but not impossible, that variations would support word size allocations.
    The VM heap might use word boundaries (but not size), where the application heap certainly does (word boundary.)My understanding of it is that the application would request blocks from the OS, and then something like malloc would manage the memory within the allocated blocks. malloc (or whatever equivalent Java uses) would have to keep track of the memory it has allocated somehow, and I would think it would have to do this at the word level, since it's most commonly going to allocate memory at the word level to be references to other objects, etc.
    So I guess my question here would really be, if the dynamic memory heap is so named because there has been a memory management strategy that relied upon a heap data structure (which I've found no proof, but have found some suggestive literature), then would that probably have applied at the OS Page Fault level, tracking allocated blocks, or would that have applied at the malloc level, allocating individual words as necessary?
    >
    4. What would be the most likely property one would use as a key. That is, what makes the root item on the heap ideal?"Key" is not a term that will apply in this discussion.
    You appear to be referring to strategies for effective allocation of memory such as allocations from different regions by size comparison.
    It is possible that all levels might use such an allocator. General purpose applications do not sort allocations though (as per your one reference that mentions 'key'.) Sorry, I got the term "key" from an article I read regarding heaps, that indicates that a "key" is used to sort the elements, which I guess would be a more generalized way to make a heap than assuming a natural ordering on the elements in the heap. I'm not sure if the terminology is standard.
    >
    5. Would a industrial strength system like the jvm use a (perhaps modified or tuned) heap for this sort of task, or would this typically be too naive for an real world solution today?Again too indefinite. The Sun VM uses a rather complicated allocator, the model for which originated after years of proceeding research certainly in Smalltalk and in Lisp as well, both commercially and academically.
    I am sure the default is rules driven either explicitly or implicitly as well. So it is self tuning.
    There are command line options that allow you to change how it works as well.I guess perhaps I could attempt to clarify my initial question a bit.
    There is a 1:1 correspondence between the runtime stack, and a stack data structure. That is, when you call a function, it pushes a stack frame onto the runtime stack. When you return from a function, it pops a stack frame from the runtime stack. This is almost certainly the reasons the runtime stack is named as it is.
    The question is, is there or has there ever been a 1:1 correspondence between some aspect of the dynamic memory heap or how it is managed, and a heap data structure? If so, it would explain the name, but I'm a bit puzzled as to how a heap data structure would be of assistance in creating or managing the dynamic memory heap. If not, on the other hand, then does anybody know where the name "heap" came from, as it applies to the dynamic memory pool?
    A.

  • How to manage connections in production environment

    Hai
    We are using Statless Session beans which will call bean managed entity beans
    as a developer we completed the first phase of our project successful but when
    comming to the production environment they(Higher authoritites of the company)
    are asking abt the connection pools. Suppose if my wlserver5.1 has 100 hits per
    a second how many connection pools i need to put in the production envirornment
    and also is it the right way to increase the connectionpools depends upon the
    clients hits????. Can anybody help me regarding this It is very urgent for us.
    Thanks in advance
    Cheers Kiran Nallam

    Hi,
    It all depends on your infrastructure and what products you are using.
    In most cases dev/prod are on different servers and are separate installs, this is the route I would promote.
    Some clients do share an environment for dev/test but not all, not everybody is the same.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Production Environment Definition

    Hello again,
    We are just starting to look at what it will take to setup a production
    Forte environment. I have some general questions regarding
    considerations that may affect the environment definition and thought
    maybe some of the more experienced users could share some thoughts on the
    following:
    1) What factors lead to the creation of multiple production environments?
    a. How many environments should you use in a production situation?
    b. Do people create separate environments for separate business units?
    c. Are there performance improvements to be had by restricting the
    number of server and client nodes included in a single environment?
    d. How do the performance benefits of multiple environments compare to
    the additional complexity of managing and maintaining multiple connected
    environments?
    The initial need is for an environment that will service approximately 50
    clients and contain a couple of server nodes (database and service
    related). However, as the environment grows, it could easily grow to a
    size of 600 clients encompassing approximately 15-20 server nodes.
    At this point in time, there is no need for the failover support of
    connected environments, but this is something we will need to add as the
    environment absorbs applications with high reliability needs. Should the
    environments be setup and connected right away or can this be easily
    added on an "as needed" basis? What other recommendations would you
    make?
    Has anyone taken advantage of Forte consulting services in defining the
    production environment? Where you satisfied with the results of the
    service?
    Thanks.
    Bradley Wells
    [email protected]
    Strong Capital Management, Inc
    http://www.strong-funds.com

    On Tue, 10 Feb 98 13:52:00 PST Brad Wells <[email protected]>
    writes:
    At this point in time, there is no need for the failover support of
    connected environments, but this is something we will need to add as
    the
    environment absorbs applications with high reliability needs. Should
    the
    environments be setup and connected right away or can this be easily
    added on an "as needed" basis? What other recommendations would you
    make?
    From the Forte Systems Management point of view, you can add them "asneeded"
    fairly easily.
    Now from the application source code point of view, implementing
    Fail/Over support
    is a different story... You will need to check your SO's dialog
    durations, handle
    DistributedAccessExceptions, "warm-up" your distributed references for
    F/O,
    design a solution for restoring global transient data, do lots of
    testings etc...
    So implementing Fail/Over is not only related to systems-management
    issues, it can
    have some influence on your application(s) source code.
    Hope this helps,
    Vincent Figari
    You don't need to buy Internet access to use free Internet e-mail.
    Get completely free e-mail from Juno at http://www.juno.com
    Or call Juno at (800) 654-JUNO [654-5866]

Maybe you are looking for

  • Problem with Printer Margins

    I am working from a Dell Optiplex 755 with Windows XP, Illustrator CS2, Acrobat 8 Pro and an HP 4345 printer. The problem is summarized in the below paragraphed points: We print on an 8.5x11 sheet of paper, with artwork sized for 8.5 x 11 exactly. Th

  • Changing states from within a component

    Let's say that I have a TileList that is rendering data in a VBox. Eventually the TileList fills up and starts scrolling. I want to change states when clicking on item in the TileList. I don't want to place the click-attribute in the TileList, becaus

  • In iTunes can I make a song apart of two different albums? How?

    I have volume 4 of glee music in one album but then I have a lot of other stuff in the complete season 2 album and I want the volume 4 stuff in the season 2 album but I also want it in the volume 4 album. Can I do that? I have the Yosemite update and

  • TV Overscan option is shaded out on nvidia-settings - How to fix it?

    I had this problem before already, but I forgot what I did to able to change nvidia-setting overscan option. At the moment it is shaded out so that I can't change it. What can I do? Nevermind... I kind of solved it. I didn't have to do it for this tv

  • Transfer Stock to subcontractor with reference of Sales order

    Hi, Scenario: MTO strategy We have one operation is subcontracting. When we take MRP run all PurReq. & PLOrd are reference to sales order. But  when we want to transfer material to subcontractor through TCode MB1B Mvt type 541 that time system issue