Adjusting CALCLOCKBLOCK and data cache

if my block size is 37464 then what should my CALCLOCKBLOCK size in the cfg.file be
Per doc (Max=half of the number of block that can fit in to the data cache)
and my data cache is 3072 kb
what is the best setting for the data cache and the CALCLOCKBLOCK
or should i just use a Data memory loclking
and if i do use the the data cache lock what would be the CALCLOCK BLOCK Setting
this is how is set it up in my cfg.file
//Set Calc Lock Block, How many block can be locked at one time
CALCLOCKBLOCKHIGH 5000
CALCLOCKBLOCKDEFAULT 2000
CALCLOCKBLOCKLOW 500
Please advise
Edited by: Next Level on Jul 4, 2012 2:05 AM
Edited by: Next Level on Jul 4, 2012 2:11 AM
Edited by: Next Level on Jul 4, 2012 2:35 AM

FWIW, I have never, and I mean never touched CALCLOCKBLOCK. I can see why one might want to fiddle with it, but in 18 years (yikes!) of BSO development, I have never seen a reason to do it.
Why do you want to change that value? 99.999% of the time, the issue isn't the number of blocks in memory isn't the issue, but the ability to have the default (I think) 100 blocks in memory. I suppose one could have a calc that touches more blocks, but I have never seen it. It would be slow, for sure and I would likely find another way to do the calc.
Regards,
Cameron Lackpour

Similar Messages

  • Dynamic Calc Issue - CalcLockBlock or Data Cache Setting

    We recently started seeing an issue with a Dynamic scenario member in our UAT and DEV environments. When we tried to reference the scenario member in Financial Reports, we get the following error:
    Error executing query: The data form grid is invalid. Verify that all members selected are in Essbase. Check log for details.com.hyperion.planning.HspException
    In SmartView, if I try to reference that scenario member, I get the following:
    The dynamic calc processor cannot allocare more than [10] blocks from the heap. Either the CalcLockBlock setting is too low or the data cahce size setting is too low.
    The Dynamic calcs worked fine in both environments up until recently, and no changes were made to Essbase, so I am not sure why them stopped working.
    I tried to set the CalcLockBlock settings in the Essbase.cfg file, and increases the data cache size. When I increases the CalcLockBlock settings, I would get the same error.
    When I increased the data cache size, it would just sit there and load and load in Financial Reporting and wouldn't show the report. In SmartView, it would give me an error that it had timed out and to try to increase the NetRetry and NetDelay values.

    Thanks for the responses guys.
    NN:
    I tried to double the Index Cache Setting and the Data Cache setting, but it appears when I do that, it crashes my Essbase app. I also tried adding the DYNCALCCACHEMAXSIZE and QRYGOVEXECTIME to essbase.cfg (without the Cache settings since it is crashing), and still no luck.
    John:
    I had already had those values set on my client machine, I tried to set them on the server as well, but no luck.
    The exact error message I get after increasing the Cache settings is"Essbase Error (1042017) Network error: The client or server timed out waiting to receive data using TCP/IP. Check network connections. Increase the olap.server.netConnectTry and/or olap.server.netDealy values in teh essbase. proprieties. Restart the server and try again"
    From the app's essbase log:
    [Wed Jun 06 10:07:44 2012]Local/PropBud/Staffing/admin@Native Directory/4340/Error(1023040)
    msg from remote site [[Wed Jun 06 10:07:44 2012]CCM6000SR-HUESS/PropBud/NOIStmt/admin/Error(1042017) Network error: The client or server timed out waiting to receive data using TCP/IP. Check network connections. Increase the NetRetryCount and/or NetDelay values in the ESSBASE.CFG file. Update this file on both client and server. Restart the client and try again.]
    [Wed Jun 06 10:07:44 2012]Local/PropBud/Staffing/admin@Native Directory/4340/Error(1200467)
    Error executing formula for [Resident Days for CCPRD NOI]: status code [1042017] in function [@_XREF]
    [Wed Jun 06 10:07:44 2012]Local/PropBud/Staffing/admin@Native Directory/4340/Warning(1080014)
    Transaction [ 0x10013( 0x4fcf63b8.0xcaa30 ) ] aborted due to status [1042017].
    [Wed Jun 06 10:07:44 2012]Local/PropBud/NOIStmt/admin@Native Directory/4340/Info(1013091)
    Received Command [Process local xref/xwrite request] from user [admin@Native Directory]
    [Wed Jun 06 10:07:44 2012]Local/PropBud/NOIStmt/admin@Native Directory/4340/Info(1008108)
    Essbase Internal Logic Error [1060]
    [Wed Jun 06 10:07:44 2012]Local/PropBud/NOIStmt/admin@Native Directory/4340/Info(1008106)
    Exception error log [E:\Oracle\Middleware\user_projects\epmsystem2\diagnostics\logs\essbase\essbase_0\app\PropBud\NOIStmt\log00014.xcp] is being created...
    [Wed Jun 06 10:07:46 2012]Local/PropBud/NOIStmt/admin@Native Directory/4340/Info(1008107)
    Exception error log completed E:\Oracle\Middleware\user_projects\epmsystem2\diagnostics\logs\essbase\essbase_0\app\PropBud\NOIStmt\log00014.xcp please contact technical support and provide them with this file
    [Wed Jun 06 10:07:46 2012]Local/PropBud///4340/Info(1002089)
    RECEIVED ABNORMAL SHUTDOWN COMMAND - APPLICATION TERMINATING

  • I can't change the date on my photos. i click on adjust change and date and it does not give me the earliest date to change

    I can't change the date on my photos in events. I click on adjust date and time of selected photos and the date that I can edit is the oldest photo date and not the earliest. Help?

    iPhoto should be showing you the date of the first photo you selected.
    What do you want to achieve? It should not matter, which photo you select to adjust the time, because all photos will be corrected by the same fixed increment, if you use "Adjust Date and Time".
    If you want to set all photos to the same time, use "Photos > Batch Change > Set Date to" instead.

  • IFS and Data Cache Configuration

    Hi!!
    i installed 9iAs Enterprice Edition on my machine that already had Oracle db 8.1.7, i'm trying to set up ifs, but when i choose the option "Oracle8i is on THIS machine" in the ifs wizard i've got a protocol adapter error, if i choose the option "oracle8i on another machine" i solved the problem, why is this happening???
    the other problem it's with theb database cache i started the wizard, but in 16% when it's creating the cache it stoped it didn't show any error, i leaved for 2 days but it just reached the 16% and stayed ther.
    could you please help me..

    For iFS configuration on Enterprise Edition you can't select the option, "Oracle8i is on THIS machine". Actually this is a typo. This should be "Oracle8i is on THIS oracle_home". This will be fixed in future releases. iFS can't be configured against database cache(icache) also. So definitly you need a database which is on another oracle_home and you should select the second option, "oracle8i on another machine".
    From the second problem, I assume you are installing 9iAS on a Windows NT platform. On NT, sometimes database cache hangs. I assume it is mainly because of less memory. How much is the memory in your system? Are you running so many processes on the system while cache configuration? If so, reduce it as much as you can and retry for cache configuration. You may need to manually shutdown one SQLPLUS.EXE process in 'Task manager' started from the current cache configuration before retrying the config.

  • JMS and Data Cache Synchronization

    We have setup Toplink to do cache synchronization using Tibco JMS. For some reason Toplink connects to JMS using the name "anonymous" in addition to connecting with the user name we specify in the code. Toplink also seems to connect to some queue in addition to the Topic that is mentioned in the code. Does anyone know why Toplink would behave like this?

    would not expect TopLink to connect twice at all. Is it possible that another service from within the ApplicationServer is connecting? I would recommend changing the JNDI location or the service port of the TIBCO JMS service and altering the TopLink JMS Clustering Service URL to reflect this change. Then verify that multiple connections are still being acquired.
    --Gordon

  • Data File Cache / Data Cache

    I have a few Questions regarding Data File cache and data cache. based on teh size of the application.
    All the settings are changed by using a maxl script.
    1. Data File cache-50MB,Data Cache - 100MB
    I am using Buffered I/O, then will memory be allocated to Data File cache?
    2. It is given in DBAG that data cache & index cache should be set as small as
    possbile for both buffered & Direct I/O. The size of one of my application is
    around 11 GB. data file :11GB,index File :450MB
    I have set my index cache to 450MB and data cache to 700MB.
    Is it OK or a. what should be my data cache size?
    b. How do I calculate the optimal Data cache and index cache?
    3. the memory size of our AIX server is 6GB. If i use Directo I/O, can my sum of
    all caches be 4GB?
    4. If I use buffered I/O, according to (2), what should be my cache sizes?
    Thanks
    Amarnath

    In the DBAG it states data file cache is not used with buffered IO so the answere to 1) should be NO.
    For 2) there is a hint in the DBAG that you should check the hit ratio of the caches to verify sizing, only calculatory advice for sizing is given on the calulator cache :-( This would mean for 2b) look at the hit ration if it stays around 1.0 try to decrease it until it drops slightly. Inspect the ratios from time to time.
    3) don't know, 64bit should be no problem. But why would you do this anyway?
    Example form our settings pag total ~20GB ind ~2GB
    outline hat 11 dimensions with a block size of ~340KB largest dense ~400 members, largest sparse ~4000 members, existing blocks ~2.7 milions
    The data cache is set to 256 MB, the index cache to 64MB our hit ratios are 1.0 for index cache and 0.77 for data chache. so our data cache could be larger, but the performace of retrievals is around 3.0 seconds which is fine for our users..
    4) Check your hit ratios and try to increase or decrease the cahces in small steps (first I'd do the index cache, the if it's fine I'd tune the data cache).
    hope it helped a bit..

  • Essbase data cache setting not registering

    Hi,
    We have 64-bit essbase. I changed the data and Index cache setting (storage type Buffered IO) in essbase.
    After restarting the database, I see the value getting updated in the "Index cache current value" but the value in the "Data cache current value" is zero.
    These are planning essbase databases. The behavior is same for the native essbase dbs.
    I have also tried restarting the essbase service. The essbase version 11.1.1.3
    Please help.
    Regards,
    Avi

    Indes and data caches work differently and it is probably good to talk about that. When an Essbase database starts, it will grab all of the memory you have defined for the index cache and nothing else can get at it. Data cachs is grabbed as it is needed until it reaches the maximum. It was truly problematic in the old days when we had much less memory to work with. If you started too many applications, you took all available memory right away. Man I feel old talking about it :)

  • Essbase BSO data cache setting

    can you guys post your BSO cube compressed data size(.pag files) and data cache setting? I am trying to get a better understanding about these two areas interms of optimization. I am also curious to know the maximum data cache setting ever used on BSO cube....(1 GB is what I know as of now)
    I think the settings depends on million things (available RAM on server, performance check, etc etc) but just a quick check.
    Thanks,
    KK

    Indes and data caches work differently and it is probably good to talk about that. When an Essbase database starts, it will grab all of the memory you have defined for the index cache and nothing else can get at it. Data cachs is grabbed as it is needed until it reaches the maximum. It was truly problematic in the old days when we had much less memory to work with. If you started too many applications, you took all available memory right away. Man I feel old talking about it :)

  • Dynamic Calc processor cannot lock more than [100] ESM blocks during the calculation, please increase CalcLockBlock setting and then retry(a small data cache setting could also cause this problem, please check the data cache size setting).

    Hi,
    Our Environment is Essbase 11.1.2.2 and working on Essbase EAS and Shared Services components.One of our user tried to run the Cal Script of one Application and faced this error.
    Dynamic Calc processor cannot lock more than [100] ESM blocks during the calculation, please increase CalcLockBlock setting and then retry(a small data cache setting could also cause this problem, please check the data cache size setting).
    I have done some Google and found that we need to add something in Essbase.cfg file like below.
    1012704 Dynamic Calc processor cannot lock more than number ESM blocks during the calculation, please increase CalcLockBlock setting and then retry (a small data cache setting could also cause this problem, please check the data cache size setting).
    Possible Problems
    Analytic Services could not lock enough blocks to perform the calculation.
    Possible Solutions
    Increase the number of blocks that Analytic Services can allocate for a calculation:
    Set the maximum number of blocks that Analytic Services can allocate to at least 500. 
    If you do not have an $ARBORPATH/bin/essbase.cfg file on the server computer, create one using a text editor.
    In the essbase.cfg file on the server computer, set CALCLOCKBLOCKHIGH to 500.
    Stop and restart Analytic Server.
    Add the SET LOCKBLOCK HIGH command to the beginning of the calculation script.
    Set the data cache large enough to hold all the blocks specified in the CALCLOCKBLOCKHIGH setting. 
    Determine the block size.
    Set the data catche size.
    Actually in our Server Config file(essbase.cfg) we dont have below data  added.
    CalcLockBlockHigh 2000
    CalcLockBlockDefault 200
    CalcLockBlocklow 50
    So my doubt is if we edit the Essbase.cfg file and add the above settings and restart the services will it work?  and if so why should we change the Server config file if the problem is with one application Cal Script. Please guide me how to proceed.
    Regards,
    Naveen

    Your calculation needs to hold more blocks in memory than your current set up allows.
    From the docs (quoting so I don't have to write it, not to be a smarta***:
    CALCLOCKBLOCK specifies the number of blocks that can be fixed at each level of the SET LOCKBLOCK HIGH | DEFAULT | LOW calculation script command.
    When a block is calculated, Essbase fixes (gets addressability to) the block along with the blocks containing its children. Essbase calculates the block and then releases it along with the blocks containing its children. By default, Essbase allows up to 100 blocks to be fixed concurrently when calculating a block. This is sufficient for most database calculations. However, you may want to set a number higher than 100 if you are consolidating very large numbers of children in a formula calculation. This ensures that Essbase can fix all the required blocks when calculating a data block and that performance will not be impaired.
    Example
    If the essbase.cfg file contains the following settings:
    CALCLOCKBLOCKHIGH 500  CALCLOCKBLOCKDEFAULT 200  CALCLOCKBLOCKLOW 50 
    then you can use the following SET LOCKBLOCK setting commands in a calculation script:
    SET LOCKBLOCK HIGH; 
    means that Essbase can fix up to 500 data blocks when calculating one block.
    Support doc is saying to change your config file so those settings can be made available for any calc script to use.
    On a side note, if this was working previously and now isn't then it is worth investigating if this is simply due to standard growth or a recent change that has made an unexpected significant impact.

  • I have LR from 2 through 4.4,  I am simply trying to adjust CR2s, and convert the them JPEG.  LR says my Perfectly Clear is expired but my update is up to date and all paid for.  What is Perfectly Clear and how do i get this puppy to be part of the team?

    i have LR from 2 through 4.4,  I am simply trying to adjust CR2s, and convert the them JPEG.  LR says my Perfectly Clear is expired but my update is up to date and all paid for.  What is Perfectly Clear and how do i get this puppy to be part of the team?

    Perfectly Clear is a 3rd party commercial plugin for Lightroom.
    http://www.athentech.com/products/plugins/
    Have you ever installed it as a trial in the past?

  • BAM Report and Active data cache

    I am having problems with reports that have automatic active data retrieval.
    It does not work.
    If I load the report it extracts the data correctly from an external data source. If I Reprompt or refresh it brings back changed data from the data source.
    It does not automatically poll for changes even though it is set in the report properties.
    If I stop and start the Active data cache service it starts polling and reconnecting, but not resynching the data.
    In the Active data cache log I have the following entries :
    2005-12-01 09:36:54,062 [3576] ERROR - ActiveDataCache Viewset not found:
    2005-12-01 09:36:54,078 [3576] WARN - ActiveDataCache Exception occurred in method GetChangeList
    Stack trace:
    at Oracle.BAM.ActiveDataCache.Kernel.Viewsets.ViewsetManager.GetViewset(String strViewsetID)
    at Oracle.BAM.ActiveDataCache.Kernel.Viewsets.ViewsetManager.GetChangeList(String strViewsetID, Int32 iTimeout)
    at Oracle.BAM.ActiveDataCache.Kernel.Server.DataStoreServer.GetChangeList(String strViewsetID, Int32 iTimeout)
    Has anyone any ideas on what the problem is ?
    Thanks

    Sriram
    Can you identify your company and project.?
    a) Maximum capacity of ADC == maximum capacity of your underlying Oracle DB used by BAM as repository.
    b) Retention period - infinite - data is there in BAM ADC (also in BAM repository) till you manually delete it (there is no automatic delete of old data)
    c) Yes ADC stores its data internally in the Oracle database (aka BAM repository)
    d) No - we generally donot publish ADC internals. This is a known decision since end users will only work either to put data into ADC or plot reports/dashboards from BAM GUI. You have to be very specific on what your are trying to acheive and we can suggest possible alternates.

  • Question about MAXCPU and clearing data cache

    Hi all,
    We're running MaxDB 7.6.03.15 and have following questions. I apologize for not grouping them together with my earlier set of questions along the similar lines.
    1. What would be an appropriate value for our application for MAXCPU when running on a quad processor machine. The application consists of the MaxDB along with 4 windows services and an instance of tomcat. Can we use 4 for MAXCPU or does it always need to be something less than the total number of CPUs on the machine (less than 4 here?). Please advise.
    2. When we're doing query execution time analysis we need to be able to clear data cache in between query runs. We achieve this so far by simply bringing the database offline and then online again. The problem with this approach is its time consuming and if we do 100 query runs a lot of time is spent in waiting for the database go from online to offline to online again. Please let us know if there's an alternative faster way to achieve the same.
    Thanks for your time.
    Sincerely,
    Sameer

    > 1. What would be an appropriate value for our application for MAXCPU when running on a quad processor machine. The application consists of the MaxDB along with 4 windows services and an instance of tomcat. Can we use 4 for MAXCPU or does it always need to be something less than the total number of CPUs on the machine (less than 4 here?). Please advise.
    Of course you may set MAXCPU to 4 - but you risk that the four Usertask UKTs will use up all CPU resources on your system then.
    So it is usually better to be rather defensive with this. If you're unsure about the CPU needs of your other services better try out and check starting with MAXCPU 2.
    > 2. When we're doing query execution time analysis we need to be able to clear data cache in between query runs. We achieve this so far by simply bringing the database offline and then online again. The problem with this approach is its time consuming and if we do 100 query runs a lot of time is spent in waiting for the database go from online to offline to online again. Please let us know if there's an alternative faster way to achieve the same.
    Well, x_cons pagecache_release should do that for you.
    But I never was able to really see the effect of it (e.g. reading the pages in again).
    If I'm not mistaken in this point - which is pretty easy because the command is rarely used and sparsely documented - it just moves the pages to the lower end of the lru list and thereby making them candidates for getting moved out of cache but not actually doing it.
    However why don't you just use the resource monitor to figure out how many pages are touched by your statements?
    Then you don't have to bounce your instance.
    regards,
    Lars

  • HT1551 My appletv is properly connected to wireless network, everything looks all right but it can't adjust the time and date. As a result it doesn't work.

    My appletv is properly connected to wireless network, everything looks all right but it can't adjust the time and date. As a result it doesn't work.

    Welcome to the Apple Community.
    Assuming this is not the first time you have used your Apple TV
    You might try restarting the Apple TV by removing ALL the cables for 30 seconds.
    Also try restarting the router.
    If the problem persists, try a restore, you may want to try the previous procedures several times before doing this.
    If restoring from the Apple TV doesn't help, try restoring from iTunes using a USB cable.
    If this is a new Apple TV, in addition to trying the above, it may also be that your network router is not allowing access to the timeserver, check that your router allows access over port 123.

  • Use of calclockblock and dynamic calc cache? what is impact on essbase?

    Hi,
    Please suggest me use of calclockblock and dynamic calc cache?
    When to use which?

    Hello,
    http://docs.oracle.com/cd/E12032_01/doc/epm.921/html_techref/config/calclockblock.htm
    And here you can find the different settings on dyn calc cache:
    http://docs.oracle.com/cd/E12032_01/doc/epm.921/html_techref/config/homecfgl.htm
    good luck!

  • Data Cache and Generic Cache

    Hello,
    According to this document,
    http://download.oracle.com/docs/cd/B16240_01/doc/apirefs.102/e12639/chapter1.htm#CIAEFHFB
    I see there are two cache system in Oracle BI Server, "Data Cache" and "Generic Cache".
    What is the difference between these two kinds of cache ?
    Thanks.

    Hi user611881,
    I'll try to explain the two cache types here, but if someone can better phrase it than myself, by all means, step up.
    Data Cache:
    The OBIEE BI Server will cache the result set of SQL that is executed. These cached results are records and if OBIEE sees another report requesting the same record set that it already has in cache, it will go straight to the cache version instead of re-issuing the SQL
    Generic Cache:
    The generic cache is the presentation server cache. This has to do with the caching of particular request. If the BI Server detects a cached version of a request, it will get the request directly from the cache and not even go down to check for the cached record set.
    This is how they function. Assume you have report A with record set 1 and there's no data in either cache.
    You run Report A
    1) Presentation Service checks if the Request is in the generic cache.
    2) This check fails.
    3) Presentation service issues logical SQL query to BI Server
    4) BI Server checks if there is any record set matching the logical query in the cache
    5) This check fails.
    6) BI Server issues the physical SQL
    7) BI Server caches the result set
    8) Presentation service gets result set and formats the data
    9) Presentation service caches presentation results
    10) Report A is displayed
    That is my understanding of the Data and Generic Caches.
    Good luck and if you found this post useful, please award points!
    Best regards,
    -Joe

Maybe you are looking for

  • New 2012 R2 domain - xp clients cannot join or print

    I just migrated a 2003  domain to 2012 R2.   Things were working ok & then XP clients became AD stupid. Steps I took: Added a VM 2012 R2 DC to the domain.  Server had DNS installed.  Ran dcdiag & bpa and resolved any issues.  About a week later I mov

  • How do I multiply an image and animate it in 3d space to create a shape?

    My boss has a small clip he would like me to make. He wants me to take an image (Its just a square image) and have it start in the center of the screen, and as it scales down and moves to one of the bottom corners (I've already got that part done) th

  • SWFVISU entry for Travel workflow

    Hi Experts, I have copied the standard travel request workflow & customized it. For approval I have user decision step & all the tasks used in my workflow are custom ones. There are no std tasks. Currently when I open the workitem from UWL i see that

  • Any idea why builders are deprecated in JavaFX 8 ?

    Hi All, Greetings! I have worked with JavaFX 2.X. Now I am trying my hand in JavaFX 8 early access. I found that all the builders are deprecated in the jfx 8 build. Do anybody have any idea regarding this.? Thanks & Regards, Sai Pradeep Dandem.

  • Strange pop-up menus on any basic funcion.

    I have never seen this problem before.  On my iMac with OS X 10.8.2 I get strange pop-up menus when I attempt the most basic functions .  When I try to sign off it says New Folder.  When I attempt to close a window it says Icon and Text.  When I atte