Maximum addressabl​e memory for A31P

Hello all:
Just a quick questio, what is the max addressable memory size for an A31P? I recall somewhere some had installed 2 GB ....? That doesn't sound right, but it would be nice if it works.
TNX.
Tony.

I've replied to your PM, but let's get this out so everyone can see:
Maximum RAM for A31/p, as well as for T30 and R40 is 2GB PC2100/2700 (DDR266/333) laptop RAM. 1GB information stands from the days when these machines were released (2002) and the largest DIMM was 512MB.
Hope this helps.
Cheers,
George
In daily use: R60F, R500F, T61, T410
Collecting dust: T60
Enjoying retirement: A31p, T42p,
Non-ThinkPads: Panasonic CF-31 & CF-52, HP 8760W
Starting Thursday, 08/14/2014 I'll be away from the forums until further notice. Please do NOT send private messages since I won't be able to read them. Thank you.

Similar Messages

  • Minimum and maximum memory for starting OSB domain.

    Hi,
       How much minimum/maximum default of memory do we need to run welogic domain for OSB ? I know we can use java comand with Xmx and Xms to set amount of memory to run domain. I want to know minimum and maximum default when we start domain.
    Many thanks.

    Dear All,
    I have attach a VI. Bascially, i want to control the maximum and minimum points using the control by keying in the timestamp date and time. How do i do that/
    Thanks!
    Attachments:
    test5.txt ‏1 KB
    timestamp.vi ‏14 KB

  • Can i upgrade my memory for my hp dv7-6c60us laptop to 16G?

    i was just wondering if i can upgrade my ramm to 8g-16g ... if its even possible.. 

    Hi:
    Below is the link to the service manual for your notebook.
    http://h10032.www1.hp.com/ctg/Manual/c03099448.pdf
    Chapter one page 5 gives the supported memory allowances.
    Because you have an Intel based chipset and processor you can install 2 x 8 GB of memory for a total of 16 GB.
    Had you had an AMD based chipset and processor, the maximum would have been limited to 8 GB.
    Paul

  • Getting " Not enough memory for the operation error " in BOE

    Hi,
    We are using BOE 11.0.
    we try to schedule a new report with multiple database logins for testing in BOE. When we run the report we are getting error like "Not enough memory for the operation error"
    The same report has been executed successfully when we run using Crystal reports developer. The report is fectching less than 10K records from all the databases together.
    Please let me know which is causing the issue and also let me know is there any limit in the number of databases connecting for a single report.
    many thanks in advance for all your help.
    Cheers,
    Suri ;-)

    Hi Sarthan,
    Sorry. I'm new to BOE. We know only scheduling the reports creating folders etc. :-)
    I've seen one parameter "Maximum Cache Size Allowed (KBytes)" and the value for this parameter is 5000.
    If we change it to a big number whether we can solve the issue ? Please suggest.
    Cheers,
    Suri ;-)

  • The merge process could not allocate memory for an operation; your system may be running low on virtual memory. Restart the Merge Agent.

    I encountered this problem on our SQL2012 and I have tried different scenarios (see below) to no avail. I have decided to give up and check if someone here has encountered this and resolved it. 
    One thing I know, it's not a memory issue. Both servers we're using has lots of memory to spare and we monitor the memory as the replication goes through it's steps. 
    I hope someone can help me on this. Thanks!
    The Error:
    The merge process could not allocate memory for an operation; your system may be running low on virtual memory. Restart the Merge Agent.
    Our Scenario
    We're using SQL Server 2012 SP1. All subscriptions are pull based.
    We're using direct Merge Replication (not FTP or web sync)
    We already have 10 active replications with larger databases. Only 1 has this issue.
    Database size is less than 5 GB
    Rebuilding the publisher is not an option.
    What have I tried?
    There is no memory problem --- we have lots to spare
    I have tried re-initialization of the database. Same problem.
    I tried deleting the database and reinitializing it. Same problem
    New snapshot. Same problem.
    I tried changing the subscriber server but still same issue. 
    MCP, MCSD, MCDBA (ANXA)

    Here is the result for the sp_configure on our subscriber. We're doing a pull on the server with the issue.
    name
    minimum
    maximum
    config_value
    run_value
    max server memory (MB)
    128
    2147483647
    2147483647
    2147483647
    In addition, I made a comparison between the working servers and the one with the issue -- there seems to be a difference in the service pack. Publisher has none but the subscriber is operating on SP1. But still strange as only 1 database is affected. 
    MCP, MCSD, MCDBA (ANXA)

  • Server has not enough memory for operation (Some .rpt files not removing from Temp folder )

    We have web application developed in ASP.NET 4.0 ftramework and published on IIS. And we are using 13_0_8 version of CR.
    I am creating report files and exporting these as pdf. And I am disposing streams and report documents at the end. Initially, there wasn't any problem and temporary files which are created by CrystalReport were deleting all. But, now requests to the web application increeased to about 50.000 in a day and now some .rpt files are staying in Temp folder and I can't delete them. After recycling application pool all files are removed by IIS. Then, after 1 or 2 hours new .rpt files are creating in Temp folder. And after somewhile, application throws Server has not enough memory for operation. And, IMHO the reason is temp files. Here is the code I am using to export report as pdf.
    Questions:
    1. Is the reason of this exception is temp files in Temp folder?
    2. What is wrong in that code?
    ReportDocument report = DownloadPDF.GetReport(id);
       MemoryStream stream = (MemoryStream)report.ExportToStream(CrystalDecisions.Shared.ExportFormatType.PortableDocFormat);
       Response.ContentType = "application/pdf";
       Response.AddHeader("content-disposition", "attachment; filename=" + id+ ".pdf");
      report.Close();
      report.Dispose();
       try
       Response.BinaryWrite(stream.ToArray());
       Response.End();
       catch (Exception)
       finally
      stream.Flush();
      stream.Close();
      stream.Dispose();
    Here is the StackTrace

    Hi Farhad
    At 50,000 requests, you are more than likely running into the CR engine limit. E.g.; you're pushing way too hard... The following will be good reads for you:
    Crystal Reports 2008 Component Engine Scalability | SCN
    (The above doc does apply to current versions of CR - e.g.; no changes.)
    Crystal Reports Maximum Report Processing Jobs ... | SCN
    Scaling Crystal Reports for Visual Studio .NET
    Choosing the Right Business Objects SDK for Your Needs
    Choose the Right SDK for the Right Task
    How Can I Optimize Scalability?
    All of the above apply to your version of CR and thus the next question will be; how to proceed:
    1) Bigger, faster servers will not hurt.
    2) Web farms.
    How Do I Use Crystal Reports in a Web Farm or Web Garden?
    3) Crystal Reports Application Server, or perhaps even SAP BusinessObjects BI Platform 4.1
    Crystal Enterprise Report Application Server - Overview
    - Ludek
    Senior Support Engineer AGS Product Support, Global Support Center Canada
    Follow us on Twitter

  • Maximum no'of Objects for one class

    Hello friends
    Please, tell me....Maximum no'of Obects for One Java class.
    Thanks & Regards
    S.Rajakrishna

    Is there any limit of creating the objects for One class?Instantiated objects go into the heap. There is only one heap for the whole JVM. There is no correlation between objects in the heap and the individual class files.
    100,000 objects may or may not fit into the heap. In general they probably will. If you don't, then yes, you'll get an Out Of Memory exception.
    Regardless there is no excuse for loading 100,000 objects just to render a JSP! You don't imagine your user is going read all of them do you? Restrict them to the number of objects that the user is actually going to be able to handle at any given time.
    You can do this by putting WHERE clauses on the hibernate query, and/or by setting limits (setFirstResult and setMaxResult) on the Query object before listing it.

  • What is maximum amount of memory that oracle db can utilize ?

    Hi
    Thank you for reading my post
    What is maximum number of memory that oracle database can utilize for each of its memory structure like sga, caches....
    Thanks

    ps, the following may help too
    VLM) Configurations
    Oracle Database for Windows supports Very Large Memory (VLM) configurations in Windows 2000, Windows 2003, and Windows XP, which allows Oracle Database to access more than the 4 gigabyte (GB) of RAM traditionally available to Windows applications.
    Note:
    This feature is available on Windows 2000, Windows 2003, and Windows XP only with Intel Pentium II and above processors.
    Specifically, Oracle Database uses Address Windowing Extensions (AWE) built into Windows 2000, Windows 2003, and Windows XP to access more than 4 GB of RAM.
    The requirements for taking advantage of this support are:
    1. The computer on which Oracle Database is installed must have more than 4 GB of memory.
    2. The operating system must be configured to take advantage of Physical Address Extensions (PAE) by adding the /PAE switch in boot.ini. See Microsoft Knowledge Base article Q268363 for instructions on modifying boot.ini to enable PAE.
    3. It is advisable (though not necessary) to enable 4GT support by adding the /3GB parameter in boot.ini. See Microsoft Knowledge Base article Q171793 for additional requirements and instructions on modifying boot.ini to enable 4GT.
    4. The user account under which Oracle Database runs (typically the LocalSystem account), must have the "Lock memory pages" Windows 2000 and Windows XP privilege.
    5. USE_INDIRECT_DATA_BUFFERS=TRUE must be present in the initialization parameter file for the database instance that will use VLM support. If this parameter is not set, then Oracle Database 10g Release 1 (10.1) or later behaves in exactly the same way as previous releases.
    6. Initialization parameters DB_BLOCK_BUFFERS and DB_BLOCK_SIZE must be set to values you have chosen for Oracle Database.
    Note:
    The total number of bytes of database buffers (that is, DB_BLOCK_BUFFERS multiplied by DB_BLOCK_SIZE) is no longer limited to 3 GB.
    Dynamic SGA and multiple block size are not supported with VLM. When VLM is enabled, the following new buffer cache parameters are not supported:
    o DB_CACHE_SIZE
    o DB_2K_CACHE_SIZE
    o DB_4K_CACHE_SIZE
    o DB_8K_CACHE_SIZE
    o DB_16K_CACHE_SIZE
    o DB_32K_CACHE_SIZE
    To select the block size for the instance, use the initialization parameter DB_BLOCK_SIZE. The buffer cache size is set by the initialization parameter DB_BLOCK_BUFFERS.
    7. Registry parameter AWE_WINDOW_MEMORY must be created and set in the appropriate key for your Oracle home. This parameter is specified in bytes and has a default value of 1 GB. AWE_WINDOW_MEMORY tells Oracle Database how much of its 3 GB address space to reserve for mapping in database buffers.
    This memory comes from the 3 GB virtual address space in Oracle Database, so its value must be less than 3 GB. Setting this parameter to a large value has the effect of using more of the address space for buffers and using less AWE memory for buffers. However, since accessing AWE buffers is somewhat slower than accessing virtual address space buffers, Oracle recommends that you tune these parameters to be as large as possible without adversely limiting database operations.
    In general, the higher AWE_WINDOW_MEMORY is set, the fewer connections and memory allocations will be possible for Oracle Database. The lower AWE_WINDOW_MEMORY is set, the lower the performance.
    8. Once this parameter is set, Oracle Database can be started and will function exactly the same as before except that more database buffers are available to the instance. In addition, disk I/O may be reduced because more Oracle Database data blocks can be cached in the System Global Area (SGA).
    Note:
    Registry parameter VLM_BUFFER_MEMORY, which enabled VLM configurations in earlier releases, is not supported in Oracle Database 10g Release 1 (10.1) or later.
    VLM Instance Tuning
    VLM configurations improve database performance by caching more database buffers in memory. This reduces disk I/O compared to configurations without VLM. VLM support in Oracle Database 10g Release 1 (10.1) or later has been re-written to integrate very closely with Windows. Compared to Oracle8i release 2 (8.1.6), VLM users should see better performance with the newer implementation.
    Tuning for VLM is no different than tuning for configurations without VLM. It is an iterative task that begins by selecting appropriate DB_BLOCK_SIZE and DB_BLOCK_BUFFERS initialization parameters for the application being supported.
    Note:
    Oracle Database 10g Release 1 (10.1) or later VLM configurations do not support multiple database block sizes.
    AWE_WINDOW_MEMORY, a new registry parameter specific to VLM, tells Oracle Database how much of its address space to reserve for mapping in database buffers. It defaults to a value of 1 GB, which should be suitable for most installations. If DB_BLOCK_SIZE is large, however, the default AWE_WINDOW_MEMORY value of 1 GB may not be sufficient to start the database.
    Increasing the value of AWE_WINDOW_MEMORY will improve performance, but it will also limit the amount of memory available for other Oracle Database threads (like foreground threads). Clients may see "out of memory" errors if this value is set too large. As a general guideline, increase the AWE_WINDOW_MEMORY registry value by 20 percent.
    For example, if DB_BLOCK_SIZE is set to 8 KB, AWE_WINDOW_MEMORY is set to 1 GB, and the number of LRU latches is set to 32 (16 processor computer), then database startup fails with out of memory errors 27102 and 34. Increasing the value of AWE_WINDOW_MEMORY to 1.2 GB fixes the problem.
    Having a large cache in a VLM configuration may also slow down database writer (DBWR) threads. Having more DBWR threads will distribute work required to identify and write buffers to disk and will distribute I/O loads among threads. Initialization parameter DB_WRITER_PROCESSES enables you to configure multiple database writer threads.
    A large cache can also introduce contention on the LRU (least recently used) latch. On symmetric multiprocessor (SMP) systems, Oracle Database sets the number of LRU latches to a value equal to one half the number of processors on the system. You can reduce contention on such configurations by increasing the number of LRU latches to twice (or four times) the number of processors on the system.
    See Also:
    Oracle Database Performance Tuning Guide for more information on instance tuning
    Windows 4 GB RAM Tuning (4GT)
    The following Windows operating systems include a feature called 4 GB RAM Tuning (4GT):
    · Windows Server 2003
    · Windows 2000 Advanced Server
    · Windows 2000 Datacenter Server
    This feature allows memory-intensive applications running on Oracle Database Enterprise Edition to access up to 3 GB of memory, as opposed to the standard 2 GB in previous operating system versions. 4GT provides a tremendous benefit: 50 percent more memory is available for database use, increasing SGA sizes or connection counts.
    Large User Populations
    Several features allow Oracle Database to support an increasingly large number of database connections on Windows:
    · Oracle Database Shared Server Process, which limits the number of threads needed in the Oracle Database process, supports over 10,000 simultaneous connections to a single database instance.
    · Oracle Net multiplexing and connection pooling features allow a large configuration to connect more users to a single database instance.
    · Oracle Real Application Clusters raises connection counts dramatically by allowing multiple server computers to access the same database files, increasing the number of user connections by tens of thousands, as well as increasing throughput.
    rgds
    alan

  • Which "Optimize Memory For" setting is best ?

    I don't understand this..Or I think I've used it wrong..please help?
    By default, Premiere Pro renders video using the maximum number of available processors, up to 16. However, some sequences, such as those containing high-resolution source video or still images, require large amounts of memory for the simultaneous rendering of multiple frames. These sequences can force Premiere Pro to cancel rendering and to give a Low Memory Warning alert. In these cases, you can maximize the available memory by changing the rendering optimization preference from Performance to Memory. Change this preference back to Performance when rendering no longer requires memory optimization.
    Select Edit > Preferences, and select Memory in the Preferences dialog box.
    In the drop-down list next to Optimize Rendering For, select Memory.
    Click OK, close Premiere Pro, and reopen the project for the new preference to take effect.
    Does setting to MEMORY only benefit lower end systems with bottlenecks, or do you use this settting on higher end machines too?
    I've always had mine set to MEMORY because I thought it would make better use of the memory. Now I have a higher end machine with 16GB of ram. And I just read a thread where someone said set it to peroformance if you want to see more CPU/RAM utilization..so now I'm confused.... Do I set it to PERFORMANCE or MEMORY w/ 16GB of ram (core i7 2600k @ 4.2ghz)? Also if you could explain why. Thanks!

    Why would low-end RAM affect the Performance Setting (vs. Memory).  It seems like the Memory setting would rely more heavily on RAM (where I have no problems).
    For what it's worth, this is a dedicated "high end" editing machine, and the RAM was top performance/top dollar when purchased.
    "I believe setting the optimise to Memory will slow down the render speed to acceptable levels"  I would like to know if this is in fact really the case (which would suggest optimize to Memory is a "slower" setting than optimize to Memory).
    I am seeing this problem in CS6 with only HD footage after editing an entire feature with CS4 in 4K (.r3d) without any issues.  Seems very weird.

  • What is the maximum number of row for a table control in LabWindows/CVI ?

    I use LabWindows CVI 8.5.1 (MMI first developped in version 6.0).
    In one of our many MMI, a table control contains a list of aircraft parameters.
    We can add as many parameters (row) as we want but over 40 000 we observe a crash of the LabWindows CVI runtime.
    Our client want to inscrease the number of parameters (row) up to 200 000 !!!
    So my questions are:
    What is the real maximum number of row for the table control ?
    Is this maximum number of row different on LabWindows 2010 version ?
    Or is there an other solution ?
    Thanks

    Greetings,
    Can you clarify what you mean by "crash"? Is there an error message thrown? Is it possible that you've consumed all of the available memory?
    Please let me know.
    Thanks, James Duvall
    Product Support Engineer
    National Instruments

  • Maximum number of dimensions for an array

    What is the maximum number of dimensions an array can accept:
    a. 266
    b. 255
    c. 256
    d. 250
    which option is correct and why?

    sabre150 wrote:
    I don't know and I most definitely don't care since anything more than 2 is, in my opinion, just silly.It's also quite untractable from a memory standpoint. Say you have an 10 by 10 by 10 array. That's 10^3 elements. Now say you have 250 dimensions, that's 10^250 elements which is a huge number in the order of the number of atoms in the universe or so.
    Say you instead have just 1 element in each dimension. Then you got a total of 1^250 elements, which is 1 element. That's okay but you don't need the array because 1 element is a variable. This means an array has at least 2 elements in each dimension. That's 2^250 elements which is like 10^75 which is also a huge number. You wouldn't get that kind of memory even if you added up all computers that ever existed.
    So array dimensions in the order of 250 aren't just silly, there isn't enougth memory for them.

  • What kind of memory for my Macbook Pro Laptop

    Hello:
    I am thinking of getting more memory for my laptop.
    I am unsure on what brand to buy, so I appreciate any input you might have. I have 2 memory slots that have 2gb each. How much can I put in each slot? 4gb, 8gb?
    What kind of prices for the memory?
    Here are the specs on my laptop.
    MacBook Pro
    13-inch, Early 2011
    Processor  2.3 GHz Intel Core i5
    Memory  4 GB 1333 MHz DDR3
    Graphics  Intel HD Graphics 3000 384 MB
    Thank you!

    You can put an 8GB module in each slot, giving you a total of 16GB. But that's the maximum.
    Good sources of RAM for your model (which use 204-pin PC-10600 (1333 MHz) DDR3 SO-DIMMs) are Crucial, OWC and Corsair. Other people may recommend other brands, but I would get one of the three (and I only include Corsair because that's what I'm using).
    Good luck,
    Clinton

  • Ive deleted contacts from my contacts list but they still pop up when i go to text and type in the name of a current contact. .  How can i delete them off my phones memory for good?

    Ive deleted contacts from my contacts list but they still pop up when i go to text and type in the name of a current contact. .  How can i delete them off my phones memory for good? iPhone 4S

    At the present time, the only way to clear that is to restore the phone as new.  Perhaps a future iOS update will give us the option to clear that cache.

  • Memory full.Not enough memory for operation.

    To the Report Design Team,
    It has been suggested that I post this issue here based on another thread I started: Re: Memory full.Not enough memory for operation.
    The template contains a single 24 bit PNG image as a background in the Page Header section, plus 9 image records from a SQL Server image field over 3 pages (3 images per page).  The images are all JPG at 3000x2000 (digital camera).
    To summarize, we are experiencing three problems.
    1. CR2008SP2 Designer issues a Memory Full error when Exporting to PDF.
    2. Using a 3rd party PDF Printer (Bullzip PDF) and CR2008SP2 Designer generates a PDF, however, the CR2008SP2 API throws a "Memory full" exception suggesting Designer and API are different (see stack trace in linked message above).
    3. With only 3 images on 1 page, CR2008 Designer Export to PDF takes 110 seconds to generate a 24MB PDF while our third party PDF Printer generates the same report in 17 seconds and 580KB.
    Edited by: Don Williams on Sep 30, 2009 6:43 AM

    The .rpt file size is 14MB with the Data Save option enabled, 12MB without Data Save.  Presumably the 12MB file size is because of the 24bit PNG we have as our background.
    The Designer executes the report in less than a second and we can scroll through all pages and see the image fields perfectly.
    When we Export to PDF, the Designer takes a long time, eventually gets to the 77%, the 7th record and returns "Export report failed" followed by "Memory full".  If we export only page 1 of the 3 pages, it also returns a Memory full error.  However, when the same report is run with only 1 page, that page exports to PDF but with a ridiculously large size and export time.
    The machine has 2GB of physical memory with an 8GB pagefile with Windows 2003 (latest everything).  The process runs up to about 1GB before reporting the memory full error.
    We've also tried a variety of other suggestions posted in the other thread with no success.
    We're happy to provide the RPT file to the Report Team to diagnose the problem.  Ultimately, we need to be able to produce a 15 page report with approximately 45 images.
    Our preferred scenario is fixing problem 2.  The CR Designer seems quite capable of rendering our report and printing it to our third party PDF printer in a timely manner with small size.  However, the API reports memory full.
    The API resides in a dedicated reporting web service with NO other code except for loading the report, setting parameters and printing.  When executing, it uses up to about 1.1GB before reporting the error.
    Are there any other suggestions for fixing what we have?  Are there known problems with large images in reports?  Do we need to lodge a formal support request?
    Regards,  Grant.
    PS.  Grr and my message formatting is lost when I edited this message!!!
    There is a 1500 character limit and then all formatting is removed to save space. Break you posts into separate entries.
    Edited by: grantph on Sep 30, 2009 2:49 AM

  • HT1270 My mid 2010 MacBook Pro has the maximum 8 GB memory cards installed. Can it be upgraded to 16 GB? Are there compatible 8 GB cards available?

    My mid 2010 MacBook Pro has the maximum 8 GB memory cards installed. Can it be upgraded to 16 GB? Are there compatible 8 GB cards available?

    I have a similar question..  13" macbook pro, mid 2010- having problems with spinning beach ball..  like someone from Mars has taken over the computer..   - the specs are below..   would it help if I added more RAM?   If  yes, any suggestions where/ what to buy?  This only started after upgrade to Mountain Lion from SNL. 
    MacBook Pro 13 inch mid 2010
      Model Identifier:    MacBookPro7,1
      Processor Name:    Intel Core 2 Duo
      Processor Speed:    2.66 GHz
      Number of Processors:    1
      Total Number of Cores:    2
      L2 Cache:    3 MB
    Memory  4 GB 1067 MHz DDR3
      Bus Speed:    1.07 GHz
      Boot ROM Version:    MBP71.0039.B0E
      SMC Version (system):    1.62f7
    Graphics  NVIDIA GeForce 320M 256 MB
    BANK 0/DIMM0:
      Size:    2 GB
      Type:    DDR3
      Speed:    1067 MHz
      Status:    OK
      Manufacturer:    0x80AD
      Part Number:    0x484D54313235533654465238432D47372020
      Serial Number:    0x27A0B2F9
    BANK 1/DIMM0:
      Size:    2 GB
      Type:    DDR3
      Speed:    1067 MHz
      Status:    OK
      Manufacturer:    0x80AD
      Part Number:    0x484D54313235533654465238432D47372020
      Serial Number:    0x2740B300

Maybe you are looking for

  • Startup disk full error - boots normally but won't start finder

    I'm having a problem in which my iMac when it boots gives me an error that the startup disk is full, but other than that it can boot normally and allow me to log in, but after that the startup stalls. I never get past the default "outer space" deskto

  • How to avoid song ending up in "Compilations"?

    I want to record songs directly off of a commercial CD for my own use. I inserted the CD and the list of songs came up. In iTunes preferences, I unchecked the "Group Compilations when browsing" and I chose a specific directory that the songs should g

  • Skipping header in external table...

    Is it possible to define an external table within OWB to skip the first few rows of a file? I haven't been able to find anything in the configuration of an external table that allows you to skip lines. If it is not available, will it be be added to O

  • CMNS support is not present in this image

    Hi, We are using a C1721 router to connect a X25 line to an ethernet port. We need to enable CMNS and we get the following error: GDM_Cisco#configure terminal Enter configuration commands, one per line.  End with CNTL/Z. GDM_Cisco(config)#interface f

  • Monitoring ASA in Cisco Prime infrastructure 2.2

    Hi everybody.  I have some issue with cisco Prime Infrastructure  2.2.  We use this product in our network for monitoring network devices and I need to monitor also our ASA 5550 device with soft 8.4.  I  added ASA to Cisco PI but see only basic compo