Buffer Pool Question regarding db_nk_cache_size parameter

If I set db_nk_cache_size parameter to db_16k_cache_size=50M then I create a tablespace using blocksize=16k, will all block objects in the tablespace be automatically pooled in my 16k cache?
Matt

Hi..
The values of db_keep_cache_size and db_recycle_cache_size are by default 0.You can create the RECYCLE and KEEP pools by specifying the following init.ora parameters:
DB_KEEP_CACHE_SIZE = <size of KEEP pool>
DB_RECYCLE_CACHE_SIZE = < size of RECYCLE pool>
When you use the above parameters, the total memory allocated to the block buffers is the sum of DB_KEEP_CACHE_SIZE, DB_RECYCLE_CACHE_SIZE, and DB_CACHE_SIZE.
Regards,
Anand

Similar Messages

  • KEEP and RECYCLE buffer pools question...

    About the DB buffer cache's two optional buffer pools: KEEP and RECYCLE pools.
    If i want a table to be kept in KEEP buffer pool, can i specify that in CREATE TABLE or ALTER TABLE syntax? Or is it possible only at tablespace level?

    HI,
    You can keep tables at keep pool by specifying storage (buffer_pool keep) parameter.

  • MULTIPLE BUFFER POOL의 개념 및 사용 방법 (ORACLE8)

    제품 : ORACLE SERVER
    작성날짜 : 1999-05-13
    Multiple Buffer Pool의 개념 및 사용 방법
    1. 필요성
    table이나 index 등 segment는 그 사용 빈도나 중요도 등에 따라 memory에
    buffering되는 것을 달리 할 필요가 있다. Oracle8에서는 buffer cache에 대해서
    multiple buffer pool이라는 새로운 특성의 개념을 제공하여 segment마다 다른
    buffer를 사용할 수 있도록 하고 있다.
    multiple buffer pool은 'keep', 'recycle', 그리고 'default' buffer pool로
    구성되며, 이것을 control하기 위한 internal algorithm은 하나의 buffer pool을
    사용할 때와 대부분 마찬가지다. 즉, 기존의 CACHE option이나 full table scan 시
    LRU end에 위치시키는 것 등은 모두 변함이 없으며, 단지 그러한 기법들이 각
    buffer마다 별도로 적용된다는 것 뿐이다.
    2. buffer의 종류
    multiple buffer pool의 주요 목적은 서로 다른 형태로 사용되는 것을 나누어 놓
    아 서로 방해가 되지 않도록 하는 것으로 정리할 수 있으며, 각각 다음과 같은
    경우에 사용하도록 한다.
    (1) KEEP buffer pool : 가능한 한 memory에 오랫동안 유지되어져야 하는
    segment를 위해 사용되어져야 한다. 자주 사용되어지고 cache size의 약 10%
    전후의 크기를 가진 segment가 이 pool을 사용하기에 적당하다.
    그러나, 여기에서도 Oracle7.3의 CACHE option과 마찬가지로 새로이 access
    되는 segment에 의해 LRU end 쪽으로 이동하는 것이 가능하므로 항상 cache
    된다고 보장할 수는 없다.
    적당한 크기로 지정하는 것이 중요한데 당연히, 동시에 memory에 올려지기를
    바라는 object들의 크기의 합보다는 커야 한다.
    (2) RECYCLE buffer pool : 자주 사용되어지지 않거나, buffer pool의 두배보다
    큰 정도의 큰 segment가 index search를 하는 작업 등에 사용되어지도록 한다.
    (3) DEFAULT buffer pool : 위의 두 buffer pool에 할당되지 않은 나머지는
    default buffer pool이 된다. 그러므로 KEEP이나 RECYCLE buffer pool은
    없어도 반드시 default buffer pool은 존재하게 된다.
    이 buffer pool은 Oracle7의 하나의 buffer pool과 같다.
    3. buffer pool을 설정하는 방법
    이러한 종류의 buffer pool을 지정하기 위해서 BUFFER_POOL_KEEP과
    BUFFER_POOL_RECYCLE이라는 parameter가 존재하며, DB_BLOCK_BUFFERS
    와 DB_BLOCK_LRU_LATCHES parameter도 함께 고려하여야 한다.
    syntax는 다음과 같다.
    BUFFER_POOL_KEEP=(buffers:<value>,lru_latches:<value>) 혹은
    BUFFER_POOL_KEEP=<value>
    BUFFER_POOL_RECYCLE=(buffers:<value>,lru_latches:<value>) 혹은
    BUFFER_POOL_RECYCLE=<value>
    위의 syntax에서 보는 바와 같이 각 pool에 대해서 buffer의 갯수 뿐 아니라
    LRU latch의 갯수도 지정할 수 있다. 만약 지정하지 않으면 그 pool에 대해서
    하나의 latch가 할당되는 것이다.
    DEFAULT pool에 대해서는 명시적으로 block의 갯수나 latch의 갯수를 지정할
    수 없고, 대신 전체 block의 갯수 (DB_BLOCK_BUFFERS)와 전체 LRU latch의
    갯수 (DB_BLOCK_LRU_LATCHES)에서 KEEP과 RECYCLE에 할당된 각각의
    값을 뺀 것만큼 default pool에 할당된다.
    간단한 예제로 설명하면 다음과 같다.
    예를 들어 initSID.ora file에 다음과 같이 parameter가 설정되어 있다고 가정한다.
    DB_BLOCK_BUFFERS=1000
    DB_BLOCK_LRU_LATCHES=6
    BUFFER_POOL_KEEP=(buffers:400,lru_latches:2)
    BUFFER_POOL_RECYCLE=100
    이러한 경우 KEEP pool에 대해서는 400개의 block과 2개의 LRU latch가 할당
    되고 RECYCLE pool에는 100개의 block과 1개의 LRU latch가 할당된다. 그리
    고 DEFAULT pool에는 500 (1000-400-100) 개의 block과 3 (6-2-1)개의 LRU latch
    가 할당되게 된다.
    각 LRU queue에 대해서 block은 균등하게 배분된다.
    즉, 이 예에서 DEFAULT queue는 LRU 1번이 167개의 block을 LRU2도 167개, 그리
    고 LRU3은 166개의 block을 가지게 되며, KEEP queue는 두개의 latch가 각각
    200개씩의 block을 그리고 RECYCLE queue는 100개의 block을 가지게 된다.
    이러한 정보는 v$buffer_pool을 통해 확인이 가능하며,
    이 예의 경우 다음과 같이 조회된다. 여기에서 set_count가 각 pool에 할당된
    latch의 갯수이며, lo_bnum과 hi_bnum이 buffer의 range이다.
    SQL> select * from v$buffer_pool;
    NAME      LO_SETID HI_SETID SET_COUNT BUFFERS LO_BNUM HI_BNUM
         0 0 0 0 0 0
    KEEP 4 5 2 400 0 399
    RECYCLE 6 6 1 100 400 499
    DEFAULT 1 3 3      500 500 999
    각 queue는 최소 50개의 block은 할당받아야 하며, 그렇지 않은 경우에는 오류
    가 발생한다. 즉 예를 들어, BUFFER_POOL_KEEP=(buffers:100, lru_latches:3)과
    같이 설정하면 alert.log file에 "Incorrect parameter specification for
    BUFFER_POOL_KEEP"이라는 오류 메시지가 적히게 되며, 100개의 block에 대해서
    최대 두개의 LRU latch만이 가능하게 된다.
    4. buffer pool을 지정하는 방법
    BUFFER_POOL이라는 Oracle8에서 새로 추가된 storage 절의 parameter를 이용
    하여 segment가 사용할 default pool을 지정할 수 있다. segment의 모든 block은
    지정된 pool을 사용하게 되며, 아래의 예제와 같이 사용하면 된다.
    CREATE TABLE keep_table(t NUMBER(10)) STORAGE (BUFFER_POOL KEEP);
    ALTER TABLE recycle_table storage(BUFFER_POOL RECYCLE);
    BUFFER_POOL은 tablespace나 rollback segment에 대해서는 지정할 수 없으며,
    clustered table에 대해서는 cluster level에서만 지정이 가능하다. partition
    table에 대해서는 각 partition별로 pool을 지정하는 것이 가능하다.
    일단 segments가 적당한 pool에 할당이 되고 난 후에는, logical hit ratio나
    free buffer waits와 같은 다양한 통계 정보가 확인 가능하다.
    이러한 통계 정보를 담고 있는 view는 v$buffer_pool_statistics이며,
    이 view는 $ORACLE_HOME/rdbms/admin/catperf.sql을 수행하면 생성된다.

  • Buffer pool parameter vs LRU/MRU

    Dear Friends,
    I know that in the database buffer cache , the following parameters define the sizes of the caches for buffers :
    –     Default : DB_CACHE_SIZE
    –     Keep : DB_KEEP_CACHE_SIZE
    –     Recycle: DB_RECYCLE_CACHE_SIZE
    And also I know LRU and MRU also works in this regards .Now my ques is , when the "keep pool" and " the "buffer pool" is on then LRU and MRU are active ?
    Both the above 3 pools and (LRU+MRU) is working together ? I have confusion about that . plz help .. ..

    I am getting buffer busy waits on some tables.
    Users are not experiencing any problem.Looks like you got bit by the CTD troll while sleeping.
    Note also that (if I'm reading the report alright) out of 23 mins you have 6 seconds accounted to buffer busy waits.
    Read the sample chapter here.

  • Db_keep_cache_size shows 0 when i keep object in KEEP buffer pool !

    Dear Frineds ,
    I use Oracle 10g . Form the oracle 10g documentaiton, I get the following information regarding ASMM (Automatic Shared Memory Management) :
    The following pools are manually sized components and are not affected by Automatic Shared Memory Management:
    Log buffer
    Other buffer caches (such as KEEP, RECYCLE, and other non-default block size)
    Fixed SGA and other internal allocations
    Now plz see the following examle :
    1) SQL> select sum(bytes)/1024/1024 " SGA size used in MB" from v$sgastat where name!='free memory';
    SGA size used in MB
    247.09124
    2) SQL> show parameter keep_
    NAME TYPE VALUE
    db_keep_cache_size big integer 0 (Here db_keep_cache_size is 0 )
    3) Now I keep the scott's dept table to KEEP cache :
    SQL> select owner,segment_type,segment_name,buffer_pool from dba_segments where buffer_pool != 'DEFAULT';
    no rows selected
    SQL> alter table scott.dept storage(BUFFER_POOL KEEP);
    Table altered.
    SQL> select owner,segment_type,segment_name,buffer_pool from dba_segments where buffer_pool != 'DEFAULT';
    OWNER SEGMENT_TYPE SEGMENT_NAME
    SCOTT TABLE DEPT
    4)
    After doing this , I have to see the following parameter :
    SQL> show parameter keep
    NAME TYPE VALUE
    db_keep_cache_size big integer 0
    SQL> select sum(bytes)/1024/1024 " SGA size used in MB" from v$sgastat where name!='free memory';
    SGA size used in MB
    246.76825
    Here I see that my sga is used but "db_keep_cache_size" still shows the '0' .
    Can u plz explain why this parameter value shows '0' now ?
    Thx in advance ... ...

    Hi,
    I am not sure I have understood the question fully but if you are trying to monitor usage of the buffer pools you should use some of the dynamic views like in the example query below. If this is not what you are interested in let me know.
    SELECT NAME, BLOCK_SIZE, SUM(BUFFERS)
    FROM V$BUFFER_POOL
    GROUP BY NAME, BLOCK_SIZE
    HAVING SUM(BUFFERS) > 0;

  • Why should recyle cache should be smaller then buffer pool.

    Hi,
    I am going through a documnetaion, where i read that your recyle cahche should be smaller then your buffer pools and it should be use buffers more quickly then .the idea behiend having a recycle cache is that if the sql statment is tuned then put the large segmant in recycle cahche so it does not effect others segmant in cache to edge out more frequentlly then it should be?now i wanted to know that what is the reason of taknig recyle cache smaller then buffer pool and why it should use buffers more quickly then the default buffer pool?and am i putting the right picture of recycle cache ?
    thanks and regards
    Alok Kumar...

    Hi Alok,
    I asume you mean by buffer the default buffer cache, right ?
    The only idea why a recycle buffer cache should be smaller than another cache can be for memory reasons: not to spend to much memory for cache. But I cannot image this should be a general guide all all kind of application ...
    On the other I have the impression that your picture of the recycle cache is a little bit different to my's picture:
    I only will use different caches if have very different tables and indexes (not sql stmts). Each table and index will be assigned to a cache, e.g. the default cache and the recycle cache. The goal is, as you described, to prtend small tables and indexes from ageing out of the default cache. But the access to the recycle cache is as fast as to the default cache: both are in memory. The only diference is the administration of the blocks into the cache. The recycle cache has an administration of the blocks much easier and faster as the other one.
    I hope this will give you some answers (and porb. some new questions) ...
    greets,
    Rainer

  • Questions regarding creation of vendor in different purchase organisation

    Hi abap gurus .
    i have few questions regarding data transfers .
    1) while creating vendor , vendor is specific to company code and vendor can be present in different purchasing organisations within the same company code if the purchasing organisation is present at plant level .my client has vendor in different purchasing org. how the handle the above situatuion .
    2) i had few error records while uploading MM01 , how to download error records , i was using lsmw with predefined programmes .
    3) For few applications there are no predefined programmes , no i will have to chose either predefined BAPI or IDOCS . which is better to go with . i found that BAPI and IDOCS have same predefined structures , so what is the difference between both of them  .

    Hi,
    1. Create a BDC program with Pur orgn as a Parameter on the selection screen
        so run the same BDC program for different Put organisations so that the vendors
        are created in different Pur orgns.
    2. Check the Action Log in LSMW and see
    3.see the doc
    BAPI - BAPIs (Business Application Programming Interfaces) are the standard SAP interfaces. They play an important role in the technical integration and in the exchange of business data between SAP components, and between SAP and non-SAP components. BAPIs enable you to integrate these components and are therefore an important part of developing integration scenarios where multiple components are connected to each other, either on a local network or on the Internet.
    BAPIs allow integration at the business level, not the technical level. This provides for greater stability of the linkage and independence from the underlying communication technology.
    LSMW- No ABAP effort are required for the SAP data migration. However, effort are required to map the data into the structure according to the pre-determined format as specified by the pre-written ABAP upload program of the LSMW.
    The Legacy System Migration Workbench (LSMW) is a tool recommended by SAP that you can use to transfer data once only or periodically from legacy systems into an R/3 System.
    More and more medium-sized firms are implementing SAP solutions, and many of them have their legacy data in desktop programs. In this case, the data is exported in a format that can be read by PC spreadsheet systems. As a result, the data transfer is mere child's play: Simply enter the field names in the first line of the table, and the LSM Workbench's import routine automatically generates the input file for your conversion program.
    The LSM Workbench lets you check the data for migration against the current settings of your customizing. The check is performed after the data migration, but before the update in your database.
    So although it was designed for uploading of legacy data it is not restricted to this use.
    We use it for mass changes, i.e. uploading new/replacement data and it is great, but there are limits on its functionality, depending on the complexity of the transaction you are trying to replicate.
    The SAP transaction code is 'LSMW' for SAP version 4.6x.
    Check your procedure using this Links.
    BAPI with LSMW
    http://esnips.com/doc/ef04c89f-f3a2-473c-beee-6db5bb3dbb0e/LSMW-with-BAPI
    For document on using BAPI with LSMW, I suggest you to visit:
    http://www.****************/Tutorials/LSMW/BAPIinLSMW/BL1.htm
    http://esnips.com/doc/1cd73c19-4263-42a4-9d6f-ac5487b0ebcb/LSMW-with-Idocs.ppt
    http://esnips.com/doc/ef04c89f-f3a2-473c-beee-6db5bb3dbb0e/LSMW-with-BAPI.ppt
    <b>Reward points for useful Answers</b>
    Regards
    Anji

  • Questions regarding *dump_dest parameters and fast_recovery_area

    Hello,
    I just installed a fresh new 11.2.0.2 Database on Solaris 10.
    Everything straightforward on the parameter side!!! I tried custom install as well as general purpose template. When installing with DBCA, I set every parameters around DB Name in lowercase name.
    With this, questions are popping in my mind regarding some parameters after installation.
    First, %dump_dest parameters contains in path, two times the db name (ocpdb in my case):
    background_dump_dest       /u01/app/oracle/diag/rdbms/ocpdb/ocpdb/trace
    user_dump_dest                 /u01/app/oracle/diag/rdbms/ocpdb/ocpdb/trace
    core_dump_dest                 /u01/app/oracle/diag/rdbms/ocpdb/ocpdb/cdumpIs it normal to have ..../rdbms/dbname/dbname/..... as path, with dbname/dbname ??? Why?
    Second, the question regarding the directory structure under fast_recovery_area (new term for flash_recovery_area). The directory structure:
    oracle@enalab13:/u01/app/oracle/fast_recovery_area$ ls -l
    total 2
    drwxr-x--- 2 oracle oinstall 512 2010-10-28 19:53 ocpdb
    drwxr----- 5 oracle oinstall 512 2010-10-29 07:44 OCPDB
    oracle@enalab13:/u01/app/oracle/fast_recovery_area$ ls -l ocpdb
    total 9528
    -rw-r----- 1 oracle oinstall 9748480 2010-10-31 21:09 control02.ctl
    oracle@enalab13:/u01/app/oracle/fast_recovery_area$ ls -l OCPDB/
    total 3
    drwxr----- 5 oracle oinstall 512 2010-10-31 03:48 archivelog
    drwxr----- 3 oracle oinstall 512 2010-10-29 07:44 autobackup
    drwxr----- 3 oracle oinstall 512 2010-10-29 07:43 backupsetWhy am I having a subdir with dbname in uppercase AND in lowercase? Should I specify dbname in uppercase at db creation to have all files under the same directory, or in lowercase? Or, is it normal?
    I want to know how to do it well before reinstalling a fresh database.
    Thanks
    Bruno
    Edited by: blavoie on Oct 31, 2010 6:18 PM
    Edited by: blavoie on Oct 31, 2010 6:20 PM

    Hi,
    I just reinstalled all from scratch, everything in lowercase as well in environment variables and dbname in dbca:
    oracle@enalab13:~$ echo $ORACLE_SID
    ocpdbFast recovery area directories, dates prove that it's my fresh install:
    oracle@enalab13:/u01/app/oracle$ ll fast_recovery_area/
    total 2
    drwxr-x--- 2 oracle oinstall 512 2010-11-02 11:06 ocpdb
    drwxr-x--- 4 oracle oinstall 512 2010-11-02 11:24 OCPDB
    oracle@enalab13:/u01/app/oracle$ ll -R fast_recovery_area/
    fast_recovery_area/:
    total 2
    drwxr-x--- 2 oracle oinstall 512 2010-11-02 11:06 ocpdb
    drwxr-x--- 4 oracle oinstall 512 2010-11-02 11:24 OCPDB
    fast_recovery_area/ocpdb:
    total 9528
    -rw-r----- 1 oracle oinstall 9748480 2010-11-02 11:34 control02.ctl
    fast_recovery_area/OCPDB:
    total 2
    drwxr-x--- 3 oracle oinstall 512 2010-11-02 11:24 archivelog
    drwxr-x--- 2 oracle oinstall 512 2010-11-02 11:06 onlinelog
    fast_recovery_area/OCPDB/archivelog:
    total 1
    drwxr-x--- 2 oracle oinstall 512 2010-11-02 11:24 2010_11_02
    fast_recovery_area/OCPDB/archivelog/2010_11_02:
    total 47032
    -rw-r----- 1 oracle oinstall 48123392 2010-11-02 11:24 o1_mf_1_5_6f0c9pnh_.arc
    fast_recovery_area/OCPDB/onlinelog:
    total 0Some interresting output asked earlier in post:
    SQL> archive log list
    Database log mode              Archive Mode
    Automatic archival             Enabled
    Archive destination            USE_DB_RECOVERY_FILE_DEST
    Oldest online log sequence     4
    Next log sequence to archive   6
    Current log sequence           6
    SQL> show parameter recovery
    NAME                                 TYPE        VALUE
    db_recovery_file_dest                string      /u01/app/oracle/fast_recovery_area
    db_recovery_file_dest_size           big integer 4032M
    recovery_parallelism                 integer     0
    SQL> show parameter control_files
    NAME                                 TYPE        VALUE
    control_files                        string      /u01/app/oracle/oradata/ocpdb/control01.ctl,
                                                         /u01/app/oracle/fast_recovery_area/ocpdb/control02.ctl
    SQL> show parameter instance_name
    NAME                                 TYPE        VALUE
    instance_name                        string      ocpdb
    SQL> show parameter db_name
    NAME                                 TYPE        VALUE
    db_name                              string      ocpdb
    SQL> show parameter log_archive_dest_1
    NAME                                 TYPE        VALUE
    log_archive_dest_1                   string
    log_archive_dest_10                  string
    log_archive_dest_11                  string
    log_archive_dest_12                  string
    log_archive_dest_13                  string
    log_archive_dest_14                  string
    log_archive_dest_15                  string
    log_archive_dest_16                  string
    log_archive_dest_17                  string
    log_archive_dest_18                  string
    log_archive_dest_19                  string
    SQL> show parameter %dump_dest 
    NAME                                 TYPE        VALUE
    background_dump_dest                 string      /u01/app/oracle/diag/rdbms/ocpdb/ocpdb/trace
    core_dump_dest                       string      /u01/app/oracle/diag/rdbms/ocpdb/ocpdb/cdump
    user_dump_dest                       string      /u01/app/oracle/diag/rdbms/ocpdb/ocpdb/traceI think, next time, I'll install everything regarding oracle SID in upper case...
    Maybe it's details that I don't need to care about... I seems that something is happening bad with the management of fast_recovery_area...
    Thanks
    Bruno

  • Questions regarding Optimizing formulas in IP

    Dear all,
    This weekend I had a look at the webinar on Tips and Tricks for Implementing and Optimizing Formulas in IP.
    I’m currently working on an IP-implementation and encounter the following when getting more in-depth.
    I’d appreciate very much if you could comment on the questions below.
    <b>1.)</b> I have a question regarding optimization 3 (slide 43) about Conditions:
    ‘If the condition is equal to the filter restriction, then the condition can be removed’.
    I agree fully on this, but have a question on using the Planning Function (PF) in combination with a query as DataProvider.
    In my query I have a filter in the Characteristic restriction.
    It contains variables on fiscal year, version. These only allow single value entry.
    The DataProvider acts as filter for my PF. So I’d suppose I don’t need a condition for my PF since it is narrowed down on fiscal year and version by my query.
    <b>a.) Question: Is that correct?</b>
    I just one to make sure that I don’t get to many records for my PF as input. <u>How detrimental for performance is it to use conditions anyway?</u>
    <b>2.)</b> I read in training BW370 (IP-training) that a PF is executed for the currently set filter (navigational state) in the query and that characteristics that are used in restricted keyfigures are ignored in the filter.
    So, if I use version in the restr. keyfig it will be ignored.
    <b>Questions:
    a.) Does this mean that the PF is executed for all versions in the system or for the versions that are in the filter of the Characteristic Restrictions and not the currently set filter?</b>
    <b>b.) I’d suppose the dataset for the PF can never be bigger than the initial dataset that is selected by the query, right?
    c.) Is the PF executed anaway against navigational state when I use filtering? If have an example where I filter on field customer thus making my dataset smaller, but executing the PF still takes the same amount of time.
    d.) And I also encounter that the PF is executed twice. A popup comes up showing messages regarding the execution. After pressing OK, it seems the PF runs again...</b>
    <b>3.)</b> If I use variables in my Planning Function I don’t want to fill in the parameter VAR_VALUE with a value. I want to use the variable which is ready for input from the selection screen of the query.
    So when I run the PF it should use the BI-variable. It’s no problem to customize this in the Modeler. But when I go into the frontend the field VAR_VALUE stays empty and needs a value.
    <b>Question:
    a.) What do I enter here? For parameter VAR_NAME I use the variable name, but what do I use for parameter VAR_VALUE?  Also the variable name?</b>
    <b>4.)</b> Question regarding optimization 6 (slide 48) about Formulas on MultiProviders:
    'If the formula is using data of only one InfoProvider but is defined on a MultiProvider, the the complete formual should be moved to the single base InfoProvider'.
    In our case we have three cubes in the MP, two realtime and one normal one. Right now we have one AggrLevel (AL) on op of the MP.
    For one formula I can use one cube so it's better to cretae another AL with the formula based on that cube.
    For another formula I need the two <u>realtime</u> cubes. This is interesting regarding the optimization statement.
    <b>Question:
    a.) Can I use the AL on the MP then or is it better to create a <u>new</u> MP with only these two cubes and create an AL on top of that. And than create the formula on the AL based on the MP with the two cubes?</b>
    This makes the architecture more complex.
    Thanks a lot in advance for your appreciated answers!
    Kind regards, Harjan
    <b></b><b></b>

    Marc,
    Some additional questions regarding locking.
    I encounter that the dataset that is locked depends on the restrictions made in the 'Characteristic Restrictions'-part of the query.
    Restrictions in the 'Default Values'-part are not taken into account. In that case all data records of the characteristic are locked.
    Q1: Is that correct?
    To give an example: Assume you restrict customer on hierarchy node in Default Values. If you want people to plan concurrently this is not possible since all customers are locked then. When customer restriction is moved to Char Restr the system only locks the specific cutomer hier node and people can plan concurrently.
    Q2: What about variables use in restricted keyfigures like variable for fy/period? Is only this fy/period locked then?
    Q3: We'd like to lock on a navigational attribute. The nav attr is put as a variable in the filter of the Characteristic Restrictions. Does the system then only lock this selection for the nav.attr? Or do I have to change my locking settings in RSPLSE?
    Then question regarding locking of data for functions:
    Assume you use the BEx Analyzer and use the query as data_provider_filter for your planning function. You use restricted keyfigures with char Version. First column contains amount for version 1 and second column contains amount for version 2.
    In the Char Restrictions you've restricted version to values '1' and '2'.
    When executing the inputready query version 1 and 2 are locked. (due to the selection in Char Restr)
    But when executing the planning function all versions are locked (*)
    Q4: True?
    Kind regards, Harjan

  • Buffer pool keep and multiple db block sizes

    I have a tablespace with 8k block size (database default) and a tablespace with 16k block size. I have db_cache_size and db_16k_cache_size set (obviously).
    Also i have buffer cache keep set in the database.
    Question: If a table is placed in a tablespace with 16k block size, and it's buffer pool is keep, does it end up in the keep pool (like tables from 8k tablsepace and keep pool set), or it ends in 16k buffer?

    You can find in the following online manual
    http://download-west.oracle.com/docs/cd/B19306_01/server.102/b14220/memory.htm#i16408

  • Question regarding Command pattern

    Hi!
    I have a question regarding the Command pattern:
    //Invoker as defined in GOF Design Patterns
    public class SomeServer {
        //Receiver as defined in GOF Design Patterns.
        private Receiver receiver;
        //Request from a network client.
        public void service(SomeRequest request) {
            Command cmd = CommandFactory.createCommand(request);
            cmd.execute();
    }The concrete command which implements the Command needs a reference to Receiver in order to execute it's operation but how is the concrete command best configured? Should I send Receiver along with the request
    as a parameter to the createCommand method or should i configure the receiver inside the CommandFactory or
    send it as a paramter to the execute method? Since SomeServer acts as both client and invoker, SomeServer "knows" about the Commands receiver. Is this a bad
    thing?
    Regards
    /Fredrik

    #!/bin/bash
    DATE=$(date '+%y-%m-%d')
    if find | grep -q $DATE ; then
    echo "OK - Backup files found"
    exit 0
    else
    echo "Critical - No Backups found today!"
    exit 2
    fi
    should work too and it's a bit shorter.
    Please remember to mark the thread as solved.

  • Question on MAXTIME parameter on queue registration

    Hi
       I have a question regarding value of parameter MAXTIME - while registering for queues in XI . My intention is to prioritize queues - XBTA, XBTG, XBTZ with the *A queue having the maximum priority and the *Z having the least. Should I give a high value in seconds for the *A and a least value for *Z queue ? I read the definition of MAXTIME - as <b>'The MAXTIME parameter determines the time in seconds that is scheduled for processing a queue'</b> .
    So, based on the above definition , should we give a higher value of MAZTIME for queues with higher priority - could somebody elaborate ?

    Hi KARTHIK MATHRUBOOTHAM ,
    These r the details Queues Prioritization:
    Queues for Prioritized Message Processing
    http://help.sap.com/saphelp_nw04/helpdata/en/04/827440c36ed562e10000000a155106/content.htm
    Queue prioritization for XI processing queues uses a built in feature of qRFC. By specifying the maximum processing time for a queue, u can define how much  processing time a specific queue will get(this works like a time-scheduling procedure, time slicing). U can use this feature to add queues for high-priority and low-priority messages(for inbound and outbound processing). These queues are then named, for example XBT1* or XBT9* for inbound EO processing  and XBTA, XBTZ for outbound EO processing.
    This function uses the message filter concept and can be configured in transaction SXMB_ADM. U can define message filters for specific message attributes such as sender/reciever service, interface and so on. A filter can also be dependent on the message size.
    This function can also be useful in situations where a queue backlog needs to be processed, due to a system error or an error on the reciever side, for example. In this case, u can use a Prioritized queue for urgent messages that must be processed immediately, without having to process the whole back-log first.
    How to prioritize XI messages on the integration server
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/20bb9649-e86e-2910-7aa9-88ed4972a5f6
    The parameter QualityOfService specifies how the Integration Engine should process a message. The following values are permitted:
    1)BE(Best Effort: synchronous processing). If QualityOfService is BE, the client is sent the final status for the processing.
    2)EO(Exactly Once: asynchronous processing with guaranteed execution exactly once). If QualityOfService is EO,processing occurs asynchronously and the client only recieves a confirmation of reciept with HTTP status '200'.
    3)EOIO(Exactly Once in Order: asynchronous processing using queues, that is, guaranteed execution exactly once and maintaining the sequence of successive messages). If QualityOfService is EOIO,processing occurs asynchronously and the client only recieves a confirmation of reciept with HTTP status '200'.
    Since u had choosen EOIO ,You must also define a queue name for EOIO:
    XI.QueueId=<QueueName>
    This queue name is used in the Integration Engine to process messages in the same sequence that they arrived in.
    This QueueName can consist of a maximum of 16 characters. If the first 8 characters contain 'SAP_ALE_', these r removed since this value is reserved by SAP and used internally.
    If a value is not specified for QualityOfService , the default value 'BE' is used.
    These r following websites which u wil find helpful:
    http://help.sap.com/saphelp_nw04/helpdata/en/0d/00453c91f37151e10000000a11402f/content.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/8f/d906d01f77fa40a4c84683c3f8326f/content.htm
    http://documentation.softwareag.com/crossvision/xie311/admin/config.htm
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/a8424011-0d01-0010-e19d-e5bd8ca52244
    cheers!
    gyanaraj
    ****Pls reward points if u find this helpful

  • Buffer pool Cache is two low .

    Dear Experts ,
    All of the sudden the producation database buffer pool cacheing had fallen down to 35 %
    select
    100*(1 - (v3.value / (v1.value + v2.value))) "Cache Hit Ratio [%]"
    from
    v$sysstat v1, v$sysstat v2, v$sysstat v3
    where
    v1.name = 'db block gets' and
    v2.name = 'consistent gets' and
    v3.name = 'physical reads';
    When i run this script i am getting '93.9% hit
    but
    select name,100 - round ((physical_reads/(db_block_gets+consistent_gets))*1
    "hit ratio" from v$buffer_pool_statistics;
    NAME hit ratio
    DEFAULT 34.87
    my hit ratio is also about 93% but sudden my database perforamce gone very low .
    please advise your valuable suggestion .

    > {quote:title=user640913 wrote:}{quote}
    > Thanks for the Response .
    > My DB version 10.2.0.4  on AIX 5.3
    >
    > there are some lending users . that cause this please help me how to find the stmt that making a troble and to solve it .
    > i am copyiing  the AWR REPORT with this mail .
    >
    > AWR REPORT LOAD PROFILE .
    >
    >                        Per Second                      PerTransaction I
    > Redo size:             89,912.03                            13,411.58
    > Logical reads:           55,554.6                             8,285.20
    > Block changes:        553.50                                 82.56
    > Physical reads:        15,842.52                              2,318.37
    > Physical writes:        20.93                                      3.12
    > User calls:              269.32                                      40.17
    > Parses:                 138.35                                      20.64
    > Hard parses:          4.13                                         0.62
    > Logons:                0.81                                          0.09
    > Executes:             356.3                                       53.23
    > ransactions:          6.70                                        0
    > [eBlocks changed per Read: 1.00 Recursive Call %: 70.83
    > Rollback per transaction %: 3.84 Rows per Sort: 103.67
    >
    Since you're licensed to run the AWR why are you bothering with something as silly as that SQL for the "buffer cache hit ratio" !
    You need to compare a "pre-problem" AWR with the current AWR reports to see if you can spot the changes that might be the significant ones, but based purely on the Load Profile that you have shown, you're not really doing much work, but you're doing it the hard way. The figures that stand out are:
    {code}
    Logical reads:           55,554.6                             8,285.20
    Physical reads:        15,842.52                              2,318.37
    {code}
    You can check the instance statistics to learn more about what type of read activity is generating these physical reads, and work out whether (a) it's sensible and (b) it's supportable on your number of disc. Having said that - take a look at the "SQL ordered by Reads" section of the report to see which SQL statements are responsible for most of the reads, get the historic plans from the AWR and then work out why the plan changed and how to produce a better plan.
    See my note below that explains why my quote of your results is tidier than your original.
    Regards
    Jonathan Lewis
    http://jonathanlewis.wordpress.com
    http://www.jlcomp.demon.co.uk
    To post code, statspack/AWR report, execution plans or trace files, start and end the section with the tag {noformat}{code}{noformat} (lowercase, curly brackets, no spaces) so that the text appears in {code}fixed format{code}.
    There is a +"Preview"+ tab at the top of the text entry panel. Use this to check what your message will look like before you post the message. If it looks a complete mess you're unlikely to get a response. (Click on the +"Plain text"+ tab if you want to edit the text to tidy it up.)                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

  • Buffer pool size

    Oracle 9.2.0.4.0
    java 1.4.2
    Hi,
    How can I set the buffer pool size in oracle 9i?
    Thanks
    JN

    DB_CACHE_SIZE
    Parameter type
    Big integer
    Syntax
    DB_CACHE_SIZE = integer [K | M | G]
    Default value
    48 MB, rounded up to the nearest granule size
    Parameter class
    Dynamic: ALTER SYSTEM
    DB_CACHE_SIZE specifies the size of the DEFAULT buffer pool for buffers with the primary block size (the block size defined by the DB_BLOCK_SIZE parameter).
    The value must be at least the size of one granule (smaller values are automatically rounded up to the granule size). A value of zero is illegal because zero is the size of the DEFAULT pool for the standard block size, which is the block size for the SYSTEM tablespace.
    [email protected]
    Joel P�rez

  • Keep buffer pool tuning

    hello,
    we have development database 10gR2
    ASSM is ON...
    and the system memory is
    SQL> sho sga
    Total System Global Area 293601280 bytes
    Fixed Size 1290232 bytes
    Variable Size 209715208 bytes
    Database Buffers 79691776 bytes
    Redo Buffers 2904064 bytes
    SQL> sho parameter sga_
    NAME TYPE VALUE
    sga_max_size big integer 280M
    sga_target big integer 280M
    SQL> select component,current_size from v$sga_dynamic_components;
    COMPONENT CURRENT_SIZE
    shared pool 171966464
    large pool 4194304
    java pool 12582912
    streams pool 20971520
    DEFAULT buffer cache 79691776
    KEEP buffer cache 0
    RECYCLE buffer cache 0
    DEFAULT 2K buffer cache 0
    DEFAULT 4K buffer cache 0
    DEFAULT 8K buffer cache 0
    DEFAULT 16K buffer cache 0
    DEFAULT 32K buffer cache 0
    ASM Buffer Cache 0
    i need to ping around 5 to 6 tables in keep buffer pool since they are frequently accessed ....so just need help on how can i properly configure the keep buffer pool sizewise...
    thank you very much

    Why do you need a keep pool?
    Why would the default buffer cache not be sufficient?
    See http://asktom.oracle.com/pls/asktom/f?p=100:11:0::::P11_QUESTION_ID:1590999000346302363
    99.999999999999% of the world does not need, want, nor desire a keep and recycle pool, they add
    to your administrative overhead and are for very special edge cases Edited by: Dom Brooks on Apr 19, 2011 12:00 PM

Maybe you are looking for