What are the relations between them?

What are the relations between the following terms?
What are included in the "Variable Size"?
Why the "Database Buffers" item is not included in the "Variable Size"?
SQL> select*from v$sga;
NAME                      VALUE
Fixed Size              1333956
Variable Size         209716540
Database Buffers      255852544
Redo Buffers            5984256Also see the following example:
SQL> SHOW SGA
Total System Global Area  638670568 bytes
Fixed Size                   456424 bytes
Variable Size             503316480 bytes
Database Buffers          134217728 bytes
Redo Buffers                 679936 bytes
SQL> SELECT * FROM v$sga;
NAME                      VALUE
Fixed Size               456424
Variable Size         503316480
Database Buffers      134217728
Redo Buffers             679936
SQL> select * from v$sgainfo;
NAME                             BYTES                  RESIZEABLE
Fixed SGA Size                   2109352                No  
Redo Buffers                     13533184               No   
Buffer Cache Size                3103784960             Yes
Shared Pool Size                 822083584              Yes 
Large Pool Size                  67108864               Yes
Java Pool Size                   134217728              Yes 
Streams Pool Size                134217728              Yes
Shared IO Pool Size              0                      Yes
Granule Size                     16777216               No
Maximum SGA Size                 4277059584             No
Startup overhead in Shared Pool  251658240              No
Free SGA Memory Available        0  

SQL> select component,current_size,user_specified_size,granule_size from v$sga_dynamic_components;
COMPONENT                                              CURRENT_SIZE USER_SPECIFIED_SIZE GRANULE_SIZE
shared pool                                               192937984                   0      4194304
large pool                                                  4194304                   0      4194304
java pool                                                  12582912                   0      4194304
streams pool                                                      0                   0      4194304
DEFAULT buffer cache                                      255852544                   0      4194304
KEEP buffer cache                                                 0                   0      4194304
RECYCLE buffer cache                                              0                   0      4194304
DEFAULT 2K buffer cache                                           0                   0      4194304
DEFAULT 4K buffer cache                                           0                   0      4194304
DEFAULT 8K buffer cache                                           0                   0      4194304
DEFAULT 16K buffer cache                                          0                   0      4194304
COMPONENT                                              CURRENT_SIZE USER_SPECIFIED_SIZE GRANULE_SIZE
DEFAULT 32K buffer cache                                          0                   0      4194304
Shared IO Pool                                                    0                   0      4194304
ASM Buffer Cache                                                  0                   0      4194304

Similar Messages

  • What are the Relations between Journalizing and IKM?

    What is the best method to use in the following scenario:
    I have about 20 source tables with large amount of data.
    I need to create interfaces that join the source tables into target tables.
    The source tables are inserted every few secondes with about hundreds to thousands rows.
    There can be a gap of few seconds between the insert of different tables that sould be joined.
    The source and target tables are on the same Oracle instance and schema.
    I want to understand the role of: 'Journalizing CDC' and 'IKM - Incremental Update' and
    how can i use it in my scenario?
    In general What are the relations between 'Journalizing' and 'IKM'?
    Should i use both of them? Or maybe it is better to deelte and insert to the target tables?
    I want to understand what is the role of 'Journalizing CDC'?
    Can 'IKM - Incremental Update' work without 'Journalizing'?
    Does 'Journalizing' need to have PK on the tables?
    What should i do if i can't put PK (there can be multiple identical rows)?
    Thanks in advance Yael

    Hi Yael,
    I will try and answer as many of your points as I can in one post :-)
    Journalizing is way of tracking only changed data in your source system, if your source tables had a date_modified you could always use this as a filter when scanning for changes rather than CDC, Log based CDC (Asynchronous in ODI, Logminer/Streams or Goldengate for example) removes the overhead of of placing a trigger on the source table to track changes but be aware that it doesnt fully remove the need to scan the source tables, in answer to you question about Primary keys, Oracle CDC with ODI will create an unconditional log group on the columns that you have defined in ODI as your PK, the PK columns are tracked by the database and presented in a Journal table (J$<source_table_name>) this Journal table is joined back to source table via a journalizing view (JV$<source_table_name>) to get the rest of the row (ie none PK columns) - So be aware that when ODI comes around to get all data in the Journalizing view (ie Inserts, Updates and Deletes) the source database performs a join back to the source table. You can negate this by specifying ALL source table columns in your PK in ODI - This forces all columns into the unconditional log group, the journal table etc. - You will need to tweak the JKM to then change the syntax sent to the database when starting the journal - I have done this in the past, using a flexfield in the datastore to toggle 'Full Column' / 'Primary Key Cols' to go into the JKM set up (there are a few Ebusiness suite tables with no primary key so we had to do this) - The only problem with this approach is that with no PK , you need to make sure you only get the 'last' update and in the right order to apply to your target tables, without so , you might process the update before the insert for example, and be out of sync.
    So JKM's provide a mechanism for 'Change data only' to be provided to ODI, if you want to handle deletes in your source table CDC is usefull (otherwise you dont capture the delete with a normal LKM / IKM set up)
    IKM Incremental update can be used with or without JKM's, its for integrating data into your target table, typically it will do a NOT EXISTS or a Minus when loading the integration table (I$<target_table_name>) to ensure you only get 'Changed' rows on the load into the target.
    user604062 wrote:
    I want to understand the role of: 'Journalizing CDC' and 'IKM - Incremental Update' and
    how can i use it in my scenario?Hopefully I have explained it above, its the type of thing you really need to play around with, and throroughly review the operator logs to see what is actually going on (I think this is a very good guide to setting it up : http://soainfrastructure.blogspot.ie/2009/02/setting-up-oracle-data-integrator-odi.html)
    In general What are the relations between 'Journalizing' and 'IKM'?JKM simply presents (only) changed data to ODI, it removes the need for you to decide 'how' to get the updates and removes the need for costly scans on the source table (full source to target table comparisons, scanning for updates based on last update date etc)
    Should i use both of them? Or maybe it is better to deelte and insert to the target tables?Delete and insert into target is fine , but ask yourself how do you identify which rows to process, inserts and updates are generally OK , to spot a delete you need to compare the table in full, target table minus source table = deleted rows , do you want to copy the whole source table every time to perform this ? Are they in the same database ?
    I want to understand what is the role of 'Journalizing CDC'?Its the ODI mechanism for configuring, starting, stopping the change data capture process in the source systems , there are different KM's for seperate technologies and a few to choose for Oracle (Triggers (Synchronous), Streams / Logminer (Asynchronous), Goldengate etc)
    Can 'IKM - Incremental Update' work without 'Journalizing'?Yes of course, Without CDC your process would look something like :
    Source target ----< LKM >---- Collection table (C$) ----<IKM>---- Integration table (I$) -----< IKM >---- Target table
    With CDC your process looks like :
    Source Journal (J$ table with JV$ view) ----< LKM >---- Collection table (C$) ----<IKM>---- Integration table (I$) -----< IKM >---- Target table
    as you can see its the same process after the source table (there is an option in the interface to enable the J$ source , the IKM step changes with CDC as you can use 'Synchronise Journal Deletes'
    Does 'Journalizing' need to have PK on the tables?Yes - at least a logical PK in the datastore, see my reply at the top for reasons why (Log Groups, joining back the J$ table to the source table etc)
    What should i do if i can't put PK (there can be multiple identical rows)? Either talk to the source system people about adding one, or be prepared to change the JKM (and maybe LKM, IKM's) , you can try putting all columns in the PK in ODI. Ask yourself this , if you have 10 identical rows in your source and target tables, and one row gets updated - how can you identify which row in the target table to update ?
    >
    Thanks in advance YaelA lot to take in, as I advised I would reccomend you get a little test area set up and also read the Oracle database documentation on CDC as it covers a lot of the theory that ODI is simply implementing.
    Hope this helps!
    Alastair

  • What are the relation between JPA and Hibernate, JPA and TopLink?

    What are the relation between JPA and Hibernate, JPA and TopLink?
    Can JPA instead of Hibernate and TopLink?

    The Java Persistence API (JPA) is the object relational mapping persistence
    standard for Java. Hibernate and TopLink provide an Open source Object-relational mapping framework for Java.
    They provide an implementation for the Java Persistence API. In my opinion, both Hibernate and TopLink provide support to JPA
    and they can also be regarded as the complementary to JPA.
    Let's wait to see other person's opinions.

  • What's ASO&BSO,what are the differences between them?

    Thanks!

    To have a gist, Essbase comes with two multi-dimensional storage options:
    Block Storage Option, the original storage type for Essbase, which permits the full set of Essbase calculation functionality and stores data accessed by sparse and dense dimensions
    Aggregate Storage Option, a more recent introduction, which stores data in quite a different way to Block Storage and is particularly suited to storing read-only data with high dimensionality.

  • What's the relation between Gain and transmitted power with coverage area?

    what's the relation between Gain and transmitted power with coverage area?
    if we want add coverage area, could we just add antenna's gain only or add transmitted power only?

    That information can be found at: http://www.cisco.com/warp/customer/cc/pd/witc/ao350ap/prodlit/agder_rg.htm

  • What is the relation between rsa5 to rsa9..

    Hi All,
    I know this is very very basic question. But i need it. Why we have install application component hierarchy before rsa5 ( exactly what is does). And what is the relation between rsa5 to LBWE. I mean to say if i want get any datasource in LBWE is it necessary to install from RSA5. And in the installation of Bussines content ( rsa 5) we r having different tabs VERSION COMPARISON, TRANSFER and SELECT DELTA what is the usage of each tab. Could you pls forward me any document about this.
    Thanks in advance...

    Hi Suresh,
    Yes we have to install business content using RSA5 before we can use it in SAP R/3.
    By means of installing business content (BC) we are changeing verion of BC component from deliveried "D" to active.
    While installing BC datasource normally we dont touch any of the tabs other than transfer datasources.
    Yes we have to install datasources in RSA5 before using them in LBWE.
    Hope this helps.
    Monica

  • What is the relation between adobe forms and web dynpro

    hi
    what is the relation between adobe forms and web dynpro

    Hi Jyothsna,
    Adobe forms are advanced to smartforms and scripts.
    Adobe forms are much easier than smartforms they are online forms.
    they can be developed online.
    Adobe forms : This is another SAP tool designed to create your own forms. The transaction for this is SFP.
    How to use them: You can create a sales order form in adobe form and send to the sales reps. Sales reps can fill this form when they are not connected to internet / SAP system. Once they connect to SAP just send an e-mail to a specific user id in outlook. You got to customize and code how do you want to process once you receive this form. This enables you to create orders even when you are not connected to the system. Hence no data loss. This is not how i used though :-). This is just my idea.
    What is the difference between Adobe forms and smart forms / scripts.
    Smart forms / scripts are used to show the data in SAP. Most likely you will use these to print / display some kind of reciepts / forms. Many companies must be using this for hard copies / ALE / EDI or to transfer data from SAP to others.
    Adobe forms are used to post data into SAP from SAP too.  So this has an additional feature compared to smartforms.
    Please check this link
    http://www.erpgenie.com/index.php?option=com_content&task=view&id=600&Itemid=77
    Web dynpro in one of the component in NETWEAVER.
    Web Dynpro is the SAP NetWeaver programming model for user interfaces and provides support when developing the Web representation of business applications. The Web Dynpro model is based on the Model View Controller paradigm, and has the following features that build on the classic dynpro model:
    ·        Clear separation of business logic and display logic
    ·        Uniform metamodel for all types of user interfaces
    ·        Execution on a number of client platforms.
    ·        Extensive platform independence of interfaces
    please check this link
    http://help.sap.com/saphelp_nw04/helpdata/en/a5/1a1e3e7181b60ae10000000a114084/content.htm
    http://help.sap.com/saphelp_nw04s/helpdata/en/77/3545415ea6f523e10000000a155106/frameset.htm
    Best regards,
    raam

  • What is the relation between G/L account and Chart of Accounts

    Hello Guru's,
                       Can you please tell me what is the relation between chart of accounts and G/L accounts.
    I know there was a list or range of account numbers in chart of accounts.
      Assets 100000 - 149999
                    current assets 100000 - 149999
                         cash and cash equivalents 100000- 118999
                             100000 - 100009
                                    100000 - petty cash.
    like this i have chart of accounts. can you please tell what are the G/L accounts.
    Thanks,
    sneha.

    Easiest way is to use the FSV - e gl account hierarchy

  • What are the differences between the following Oracle database views?

    What are the differences between the following Oracle database views(or function)? Can they be instead of each other? Can all of them be used to pull out the execution plan from the library cache?
    V$SQL_PLAN
    DBA_HIST_SQL_PLAN
    dbms_xplan.display_cursor function

    junez wrote:
    What are the differences between the following Oracle database views(or function)? Can they be instead of each other? Can all of them be used to pull out the execution plan from the library cache?
    V$SQL_PLAN
    DBA_HIST_SQL_PLAN
    dbms_xplan.display_cursor functionV$SQL_PLAN-this is dynamic perfomance view which contain execution plan information for each child cursor loaded in the library cache.
    DBA_HIST_SQL_PLAN- data dictionary view also contain execution plan information for each child cursor bun in workload repository.Actually this view captures information from V$SQL_PLAN and is used with the DBA_HIST_SQLSTAT .
    dbms_xplan.display_cursor - function displays explain plan of any cursor loaded in the cursor cache. See more information below links:
    http://youngcow.net/doc/oracle10g/server.102/b14237/dynviews_2120.htm
    http://download.oracle.com/docs/cd/B14117_01/server.101/b10755/statviews_2154.htm
    http://download.oracle.com/docs/cd/B12037_01/appdev.101/b10802/d_xplan.htm#998179

  • What are the differences between TestStand Version 1.0.2 and 1.0.3?

    I have TestStand version 1.0.2 and have a customer that uses 1.0.3. I am updating test software for that customer and need to know if I can use TestStand 1.0.2. What are the differences between 1.0.2 and 1.0.3?

    vc -
    The Whatsnew.txt file for TestStand 1.0.3 contains the following text:
    Changes from TestStand 1.0.2 to 1.0.3
    TestStand 1.0.3 includes a translation to Japanese. It also
    includes other fixes mostly related to translation issues.
    1) Configure Database Options menu item was not localized.
    2) Dismissing a modal CVI dialog brought a random window to the front.
    3) Removed text from operator interface splash screens
    4) Resized operator interface panels
    5) In the TestStand - Sequence Display.vi of the LabVIEW operator interface, the
    following control descriptions didn't use resource strings:
    - Sequence Description:
    - Current Test:
    - Number of Tests:
    6) The Engine Installation Wizard wasn't localizable.
    7) Flexible MessageBox sometimes displayed only 1-2 lines. This was seen in the
    Edit Database Options Help dialog and the Adapter Configuration dialog in the
    CVI operator interface.
    8) The buttons in the Automation Adapter Configuration dialog were not localized.
    9) Font for the Add Directory to Search Directories checkbox on Find File dialog
    was incorrect for Japanese Windows.
    10) Dialog used to browse for directory to specify the report location was not translated.
    11) Database Viewer was not localized correctly.
    Scott Richardson
    National Instruments

  • What is the relation between PAYR adn PAYRq?

    What is the relation between PAYR, PAYRQ, REGUH, REGUP tables in FI?
    Which gets updated and when?
    I was validationg data from REGUP vs PAYR but the data seems to be avilable in PAYRQ and not PAYR.
    PAYRQ probably seems to be a request that is not completed.
    but not sure if this is correct.
    Please let me know the relationships among these and also when (how) REGUP and REGUH get filled.
    Thanks,
    Ven

    HI,
    this are all the payment related table Payr refers to check number and the REGUH and REGUP refers to app run data.
    all the three will be linked based on the payment run and date.
    if only f-58 u will find in payr.
    hope it is clear
    assign points.
    With regards
    krishna

  • What is the relation between KTOPL and 0CHRT_ACCTS_ATTR? both Chart of Acct

    Hi
    If an R3 field e.g. KTOPL (in datasource 0fi_gl_4)  which is a "Chart of Accout" is extracted a part of 0fi_gl_4 to BI,
    is there the need to separately extract 0CHRT_ACCTS_ATTR & 0CHRT_ACCTS_TEXT which are also "Chart of Accout"
    What is the relation between KTOPL and 0CHRT_ACCTS_ATTR & 0CHRT_ACCTS_TEXT  in R3?
    What is the relation between KTOPL and 0CHRT_ACCTS_ATTR & 0CHRT_ACCTS_TEXT  in BI?
    Thx

    Hi,
    I get these well with examples so let me see if I get your point:
    So in the tranx data the 0fg_gl will be  a value for KTOPL  e.g. 450009 as a G/L Account number? (Values for chart of Acct should be GL Accounts, right?)
    but before this tranx data is loaded, preferrably,  0CHRT_ACCTS_ATTR and 0CHRT_ACCTS_TEXT should have been extracted to BI; and data loaded; and this will be the master data and there should have been in it:
    .._ATTR   -
    .._TEXT
    450009   -
    Petty cash Expenses
    Is my understanding exact?
    Thnx

  • What is the relation between UTXJ and JMOD, how JMOD determine in order

    Hi All,
    What is the relation between UTXJ and JMOD! 
    How JMOD determine in order! ( For TAXINJ).
    Please help me
    Regards
    raj

    Dear Raj
    Both are not related in anyway.  UTXJ is for sales tax and JMOD is for excise condition.  For UTXJ, you need to create a tax code in FTXP and assign it in VK11 / UTXJ, whereas, JMOD is an excise condition type for which, you need to maintain the relevant datas in J1ID.
    Coming to your question, how JMOD is determined in sale order, it is based on delivering plant.  This delivering plant, as already said above, you will have to maintain in J1ID with the indicator "1" under declared tab in Material Chapter ID combination.
    thanks
    G. Lakshmipathi

  • What are the difference between TopLink and TopLink Essentials...?

    What are the difference between TopLink, TopLink Essentials, EclipseLink and TopLink Essentials-GlassFish?
    What is the difference of their functions?
    Edited by: qkc on Nov 21, 2009 10:52 AM

    Difference between TopLink and ToPLink Essentials:
    TopLink Essentials are the reference implementation (RI) of JPA, is an open source effort that is licensed under the Common Development and Distribution License (CDDL) v1.0. It can be freely downloaded and used under the terms of this license agreement. This means, you can only use it for the management of persistence and orm with Java EE and Java SE. The binary distribution of TopLink consist of 2 jars:
    * toplink-essentials-agent.jar: contains Java Persistence API; XML Schemas and TopLink essentials implementation
    * toplink-essentials.jar: contains java agent class requires in a standolane Java SE application
    In addition to TopLink Essentials that makes the JPA implementation alive, Oracle offers its Oracle TopLink product (10.1.3) that contains an earlier preview binary of JPA and also offers developers additional object-relational capabilities, object-XML mapping (JAXB), non-relational mapping using Java 2.0 Connector Architecture (JCA).
    Let's wait other person to answer these questions.

  • Technically, what's the relation between FlexUnit and FB?

    I'm running into many small but annoying issues when trying to do FlexUnit testing in Flash Builder 4, like:
    * FlexUnit is open source but I can't Cltr+click on things like assertEquals() to see what's going on there
    * I visit flexunit.org, click Docs and can't find assertThat() method at all while it's available in FB's code hints (and works)
    * We now have 2 documentation sets, one in the "Using Flash Builder 4" book, one on docs.flexunit.org. Some things are documented here, some others there.
    * API docs aren't even available from Adobe, are they?
    I wonder, what is the relation between FlexUnit itself and the developers working on it (Michael and Brian primarily?) and the support in Flex 4 / Flash Builder? I am not very happy with the current state which has nothing to do with FlexUnit itself (is a very nice testing framework), I just feel that a feature that ships in the quite expensive Premium edition should have a much better integration story.
    It looks like I'm ranting (and I am a little) but mainly I'd like to understand what's going on here.
    Thanks,
    Borek

    Borek:
    I understand your pain and I am sorry. Let me answer some of your questions and provide some answers and suggestions.
    Adobe decided to include FlexUnit 4 in Flash Builder 4, which was great. However, when it comes to the way in which they did so, the way in which libraries are linked and the way to plugin works, we (the FlexUnit team) are only able to provide suggestions. We have no control over the approach and little ability to influence the process.
    So, to be even more direct in my answer, there is no formal relationship between these two teams. What features Adobe chooses to support and the workflow/integration are purely business choices being made by Adobe. FlexUnit is purely a community project without funding or support from Adobe. It continues to evolve as we add new features and integrate contributions. Adobe may choose to increase the level of integration or to adopt newer versions as we proceed, but that is solely at their discretion.
    To that end, we cannot change the way the plugin works, etc. However, we will be posting information for users interested in updating their Flash Builder version with our latest code base (and addressing at least one of your issues) shortly as we prepare for our 4.1 beta.
    * FlexUnit is open source but I can't Cltr+click on things like assertEquals() to see what's going on there
    >>Adobe didn't include the code for FlexUnit in builder, only the library. You can control-click on Flex classes because they shipped both the lib and code. If you do want this functionality, you can download our code, go into the build path/library path of the project, find the place where they link in the flexunit-core libraries and edit the source attachment. That will allow the control-click to work, etc. The code that shipped with Flash Builder is tagged as 4.0.0 in github.
    * I visit flexunit.org, click Docs and can't find assertThat() method at all while it's available in FB's code hints (and works)
    >>I will try to find out why this is not showing up and get it fixed promptly. FYI, the assertThat method is just a wrapper for the org.hamcrest.assertThat from the hamcrest library. For more info and code: http://github.com/drewbourne/hamcrest-as3
    * We now have 2 documentation sets, one in the "Using Flash Builder 4" book, one on docs.flexunit.org. Some things are documented here, some others there.
    >>I don't know what is in the "Using Flash Builder 4" book...if you can point me to a link I will check it out. I have no idea what is there (nor who wrote it) The official documentation site is docs.flexunit.org. It is the only one being maintained/added to by those involved with the project.
    * API docs aren't even available from Adobe, are they?
    >>I doubt it. They don't maintain this material and they don't include it.
    Sorry if this is less than helpful, but I wanted to at least provide you some answers and context.
    Mike

Maybe you are looking for