What is the use of Indexes in ODS&Infocube??

What is the use of Indexes in ODS&Infocube??

An Index can improve the reading performance when data is searched for values of fileds contained in the index.
The fact whether the Index is used or not depends on the database optimizer, which decides it after taking the following under consideration:
- Size of the table
- Fields in the index compared to the fields in the statement
- The quality of the index (its clustering factor)
One backdraw of Index is that they decrease writing performance, as they are maintained during the write to the table.
Hope it helps you,
Gilad

Similar Messages

  • What is the use of parllalization in loading infocube by the help of dtp

    what is the use of parllalization in loading infocube by the help of dtp

    An Index can improve the reading performance when data is searched for values of fileds contained in the index.
    The fact whether the Index is used or not depends on the database optimizer, which decides it after taking the following under consideration:
    - Size of the table
    - Fields in the index compared to the fields in the statement
    - The quality of the index (its clustering factor)
    One backdraw of Index is that they decrease writing performance, as they are maintained during the write to the table.
    Hope it helps you,
    Gilad

  • What is the use of STORNO in ODS ?

    Hi Experts,
    Let me know the detail information about STORNO In ods.
    Thanks in advance
    Madhusudhan Reddy

    Hi,
    depending on the extractor in the source system, you'll get different records with different recordmodes.
    With respect to the "STORNO" to types are relevant:
    'R' Recordmodes delivers an inverted record regarding to an existing one. The rest happens in the change log.
    'D' Records will be deleted in the ODS by the delivered key. Extractors providing these Records can't be written into cubes directly.
    Is that suffiecient or is more information required.
    Cheers
    Sven

  • Why do we create indexes for DSOs and Cubes.What is the use of it?

    Hi All,
    Can you please tell me why are indexes created for DSOs and Cubes.
    What is the use with the creation of indexes.
    Thanks,
    Sravani

    HI ,
    An index is a copy of a database table that is reduced to certain fields. This copy is always in sorted form. Sorting provides faster access to the data records of the table, for example, when using a binary search. A table has a primary index and a secondary index. The primary index consists of the key fields of the table and is automatically created in the database along with the table. You can also create further indexes on a table in the Java Dictionary. These are called secondary indexes. This is necessary if the table is frequently accessed in a way that does not take advantage of the primary index. Different indexes for the same table are distinguished from one another by a separate index name. The index name must be unique. Whether or not an index is used to access a particular table, is decided by the database system optimizer. This means that an index might improve performance only with certain database systems. You specify if the index should be used on certain database systems in the index definition. Indexes for a table are created when the table is created (provided that the table is not excluded for the database system in the index definition). If the index fields represent the primary keys of the table, that is, if they already uniquely identify each record of the table, the index is referred to as an unique index.
    they are created on DSO and cube for the performance purpose ..and reports created on them wil be also more efficent ..
    Regards,
    shikha

  • What is the use of commit work and rollback work

    hao friends,
    what is the use of commitwork and rollback work where we can use these open sql statements .
    thanks,
    anji

    hi all
    i have written the following code in my development system to change the records of my database.
    SELECT * FROM zyikostl INTO TABLE itab.
      LOOP AT itab INTO wtab.
        tabix = sy-tabix.
        DO.
          READ TABLE itab INTO ktab WITH KEY kostl_alt = wtab-kostl_neu.
          IF sy-subrc = 0.
            wtab-kostl_neu = ktab-kostl_neu.
            MODIFY itab FROM wtab INDEX tabix.
            CONTINUE.
          ELSE.
            EXIT.
          ENDIF.
        ENDDO.
      ENDLOOP.
    MODIFY zyikostl FROM TABLE itab.
    In the development system i have only few records its working fine,but when it is been moved to test system there are 2500 records in the database so the report is taking much time(more than 5 hours )
    i sthere any solution for this ....
    thanks,
    vaasu.

  • What are the functions of tables in ods

    Hi,
    what are the functions of tables in ods.
    new data table
    change log table
    active data table.
    How 0recordmode work with these 3 tables.
    and how sid will work with ods?
    can anyone give detailed explanation.
    please do not give sap help links.
    Thanks,
    madhu
    Thanks

    Hi,
    new table holds the "new" loaded data in ODS. If you start a new request to load data in your ODS, it firstgoes to table new data. Then you normally activate data either manually or by process chain. Then the daa will be copied over from new to active table. The change log table holds the delta information, what will change during activation or which records will be overwritten with new data or which data is new at all.
    With recordmode you can specify which update mode will be used or more excalty which delta mode will be used. SID is necessary for master data, because ODS will only hold the SID values of the used characteristivcs and not the characteristic values itselfs. It's part of normalization used in BI.
    For more details I recommend to look here in SDN, go to a class or read help.sap.com pages.
    Regards,
    Juergen

  • What is the use of delta in BW side when we have delta in R/3

    Hi,
    what is the use of delta in BW side when we are already using delta in R/3 side?
    Like,
    Generic extraction:-
    with different delta extractions (Time stamp, Calendar day, Numeric pointer).
    Lo-Cockpit extraction with change log method.
    Co/Pa extraction with time stamp method.
    So from this delta activation first time full data will be loaded and after that only changed data will be loaded because of these deltas in R/3 side.
    So now my question is what is the exact use of delta in BW side with ODS/DSO because data already reached to BW with delta activation right.
    Can you please explain me what is the exact use of BW delta?
    Thanks in advance.
    Points to be added.
    Regards,
    Shyam.

    The purpose of Delta Queue is same in R/3 and BW.
    In BW, Delta Queue is used to capture changes made to objects within BW. (for Eg: BW Statastics)or Update data from one BW system to another BW system (if you have multiple BW systems).
    Message was edited by:
            Praveen Vujjini

  • What is the effect of INDEXING on Cube? and on DSO?

    Hi,
    I am trying to figure out the effect of indexing on a Cube and DSO.
    1. My findings so far indicates that Indexing is a database concept but so is partition but partitioning applies in the case of a Cube. Right?
    2.What is the effect of Indexing on a cube? How is it related to the "database concept that I keep reading?
    How is it implement in BI?
    3.What is the effect of Indexing on a DSO? How is it related to the "database concept that I keep reading?
    How is it implement in BI?
    Thanks

    Hi,
    thanks for the detailed information. I will appreciate some help on these follow ups on your posting:
    On your answer to 1:
    I read that there are 2 partitions, logical and physical. It is the physical partitioning which takes place at the database level, so can I assume that when you said u201CEven DSO can be partitionedu201D you meant the physical partition? And that logical partitioning is not possible in DSOs?
    On your answer to 2:
    So in the above you suggested that indexes apply not only to cubes but also to DSO with the statement that u201C..secondary indexes can be built on DSO u2026u201D
    If so, how come there is not feature to drop and rebuild indexes for data load to DSOs?
    I will appreciate a hint on the difference between u201Cindexesu201D and u201Csecondary indexesu201D. I am ok with what u201Cindexesu201D is, is the use of u201Csecondaryu201D that threw me off.//
    On your answer to 3:
    i.
    Based on the link, there does not appear to be a simply way to drop and rebuild indexes in DSOs. It appears an ABAPer needs to be involved. Isnu2019t there a simply way as in cubes, to click on buttons for the index implementation for DSOs?
    ii.
    It suggests one must choose between B-tree and Bit map indexes, right? Is this the function of the BI consultant or as DBA? If the BI consultant should be able to do this, where is it implemented in BI?
    Can you explain what you meant by u201Ccardinality of a column is lowu201D?
    iii.
    I know how in rsa1, a Cube can be partitioned on 0calmonth and 0fisper. If this is done at the database level, then can it be done also for DSO? If so, where in BI is it done for DSOs?
    You also noted that SAP DB does not support table partitioning but Oracle does.
    I know in my environment, the BASIS consultants are always talking about Oracle so I am assuming we have Oracle and not SAP DB.
    I never heard of SAP DB, is it a substitute for Oracle in SAP BI environments?
    Is it a DB that comes directly from SAP; if so, why do companies not keep the total SAP package but prefer to use Oracle?  Is it because the SAP DB does not support table partitioning?
    Thanks

  • What is the use of these filters in atg

    Hi Guys,
    In my web.xml file i have code like this
    <filter>
            <filter-name>PageFilter</filter-name>
            <filter-class>atg.filter.dspjsp.PageFilter</filter-class>
        </filter>
        <filter>
            <filter-name>ForwardFilter</filter-name>
            <filter-class>atg.servlet.ForwardFilter</filter-class>
        </filter>
        <filter>
            <filter-name>ErrorFilter</filter-name>
            <filter-class>atg.servlet.ErrorFilter</filter-class>
        </filter>
        <filter-mapping>
            <filter-name>ErrorFilter</filter-name>
            <url-pattern>/*</url-pattern>
            <dispatcher>ERROR</dispatcher>
        </filter-mapping>
        <filter-mapping>
            <filter-name>ForwardFilter</filter-name>
            <url-pattern>/*</url-pattern>
            <dispatcher>FORWARD</dispatcher>
            <dispatcher>ERROR</dispatcher>       
        </filter-mapping>
        <filter-mapping>
            <filter-name>PageFilter</filter-name>
            <url-pattern>/*</url-pattern>
            <dispatcher>REQUEST</dispatcher>
            <dispatcher>ERROR</dispatcher>       
            <dispatcher>FORWARD</dispatcher>
        </filter-mapping>
    what is the use of these filters what will happend if i remove these filters in web.xml file
    please give me clear picture on these 3 filters and how these filters are helpful in our application.

    As the name filter suggest, All the request will go though these filters if certian condition is satisfied
    In ATG web.xml contains following filter:
    <filter>
        <filter-name>PageFilter</filter-name>
        <filter-class>atg.filter.dspjsp.PageFilter</filter-class>
    </filter>
    <filter-mapping>
        <filter-name>PageFilter</filter-name>
        <url-pattern>*.jsp</url-pattern>
    </filter-mapping>
    When there is .jsp in url pattern all the request will go through Page Filter.
    Filter is java concept.For More details about Filter, Refer http://docs.oracle.com/cd/E19502-01/819-3669/6n5sg7b0b/index.html
    In order to process Jsp request, ATG use Page filter. It converts normal HttpRequest/HttpResponse to DynamoHttpRequest/DynamoHttpResponse. Please refer http://docs.oracle.com/cd/E24152_01/Platform.10-1/ATGPlatformProgGuide/html/s0801requesthandlingwithservletpipeli01.html

  • What is the use of SYSAUX Tablespace in Oracle 10G

    Dear Experts,
    Please Tell me What is the use of SYSAUX Tablespace in Oracle 10G Because
    during Web Load Testing It is going to increase countinuosly.Although i am not using this with any tables.
    How Can I find out the what is gong on in this.
    Can I Remove this tablesspace and other non using Schemas like FLOWS_020100,MDSYS,OUTLN,DIP and TSMSYS etc.

    What is SYSAUX?
    The SYSAUX tablespace provides storage of non-sys-related tables and indexes that traditionally were placed in the SYSTEM tablespace. For example, the tables and indexes that were previously owned by the system user can now be specified for a SYSAUX tablespace. Unfortunately, Oracle still places the SCOTT schema and the other demonstration schemas in the SYSTEM tablespace. Go figure.
    The SYSAUX tablespace is specified with the CREATE DATABASE command. This is demonstrated in the example database creation script in Figure 1.1.
    CREATE DATABASE test
    MAXINSTANCES 1
    MAXLOGHISTORY 1
    MAXLOGFILES 5
    MAXLOGMEMBERS 3
    MAXDATAFILES 100
    DATAFILE '/usr/oracle/OraHome1/oradata/aultdb1/test/system01.dbf' SIZE 300M REUSE AUTOEXTEND ON NEXT 10240K MAXSIZE UNLIMITED
    EXTENT MANAGEMENT LOCAL
    SYSAUX DATAFILE '/usr/oracle/OraHome1/oradata/aultdb1/test/sysaux01.dbf' SIZE 120M REUSE AUTOEXTEND ON NEXT 10240K MAXSIZE UNLIMITED
    DEFAULT TEMPORARY TABLESPACE TEMP TEMPFILE '/usr/oracle/OraHome1/oradata/aultdb1/test/temp01.dbf' SIZE 20M REUSE AUTOEXTEND ON NEXT 640K MAXSIZE UNLIMITED EXTENT MANAGEMENT LOCAL
    UNDO TABLESPACE "UNDOTBS1" DATAFILE '/usr/oracle/OraHome1/oradata/aultdb1/test/undotbs01.dbf' SIZE 200M REUSE AUTOEXTEND ON NEXT 5120K MAXSIZE UNLIMITED
    CHARACTER SET WE8ISO8859P1
    NATIONAL CHARACTER SET AL16UTF16
    LOGFILE GROUP 1 ('/usr/oracle/OraHome1/oradata/aultdb1/test/redo01.log') SIZE 10240K,
    GROUP 2 ('/usr/oracle/OraHome1/oradata/aultdb1/test/redo02.log') SIZE 10240K,
    GROUP 3 ('/usr/oracle/OraHome1/oradata/aultdb1/test/redo03.log') SIZE 10240K
    USER SYS IDENTIFIED BY "password" USER SYSTEM IDENTIFIED BY "password";
    Oracle10g sysaux
    With the new SYSAUX tablespace, Oracle comes closer to providing all the needed tablespaces for a truly OFA-compliant database right out of the box. With just one CREATE DATABASE command we can specify the SYSTEM tablespace, the TEMPORARY tablespace, the AUXSYS tablespace, the default UNDO tablespace, and the redo logs.
    Of course, with the Oracle Managed Files option you can create an entire database with a single command, but the database created is not suitable for production use and is not OFA-compliant.
    The SYSAUX tablespace is required in all new 10g databases. Only the SYSAUX tablespace datafile location is specified. Oracle specifies the remainder of the tablespace properties including:
    online
    permanent
    read write
    extent managment local
    segment space management auto
    If a datafile is specified for the SYSTEM tablespace, then one must be specified for the SYSAUX tablespace as well. If one is not specified, then the CREATE DATABASE command will fail. The only exception is for an Oracle Managed File system.
    During any update of a database to Oracle Database 10g, a SYSAUX tablespace must be created or the upgrade will fail. The SYSAUX tablespace has the same security profile as the SYSTEM tablespace. However, loss of the SYSAUX tablespace will not result in a database crash, only the functional loss of the schemas it contains.
    Can I drop SysAux?
    DROP DATAFILE
    Read this...

  • What is the use of  keyword SCAN ABAP-SOURCE

    Hello experts,
    what is the use of scan abap-source. please explain. what is tokens, statements, levels in that.

    Hi
    <b>SCAN</b>
    This statement is for internal use only.
    It cannot be used in application programs.
    <b>SCAN ABAP-SOURCE itab1 ...TOKENS INTO itab2
                           ...STATEMENTS INTO itab3.</b>
    Parts marked with " ..." are interchangeable
    <b>Addition 1</b>
    ... FROM n1
    <b>Addition 2</b>
    ... TO   n2
    Breaks down the source code table itab1 into tokens not from start to finish, but only from line n1 to line n2.
    The additions FROM n1 and TO n2 must, in this order, follow the specification of the source code table itab1.
    When using the start specification n1, use the addition WITHOUT TRMAC to ensure that there are no unnecessary database accesses to the table TRMAC.
    The end specification n2 is treated as "soft", i.e. a statement that begins on a line <= n2, but ends only on a line > n2, is returned completely.
    If the end specification n2 is split in a chain statement, only the split part up to the next comma is returned completely, not the entire chain statement up to the next period.
    Negative line specifications are not allowed and result in a runtime error.
    A line specification of 0 amounts essentially to no specification.
    If n1 number of lines in source code table, the scanner is not called (SY-SUBRC = 2).
    If n1 > n2 and n2 > 0, the scanner is not called (SY-SUBRC = 2).
    <b>
    Addition 3</b>
    ... KEYWORDS FROM itab4
    Does not return all statements, only those specified in the key word table itab4.
    If the key word table is empty (i.e. it contains 0 lines), all the statements are selected.
    The lines of the key word table are treated as a character field.
    To select a Native-SQL-statement or a macro definition, you can specify the pseudo key words EXEC_SQL or DEFINE_MACRO. It makes no difference whether the statements EXEC or DEFINE occur as well. Native SQL statements and macro definitions are returned as one statement (of type E or M even if the expansion of a macro definition results in more than one statement.
    If the key word table contains a blank line, blank statements are also selected.
    <b>Addition 4</b>
    ... LEVELS INTO itab5
    Stores details about each edited source code unit (source code table itab1 itself, expanded include-programs, expanded macro definitions) in the level table itab5.
    Specification of a level table makes sense only with the addition WITH INCLUDES.
    The level table itab5 must have the structure SLEVEL.
    The fields of the structure SLEVEL - and consequently the columns of the level table itab5 have the following meaning:
    TYPE
    Type of source code unit with the following possible values:
    P (Program)
    D (Internal DEFINE macro)
    R (Macro from table TRMAC)
    NAME
    Name of source code unit (name of include program, macro name)
    DEPTH
    Current nesting depth of source code unit (>= 1)
    LEVEL
    Index of superior (i.e. including or calling) source code unit in the level table (>= 1, if DEPTH >= 2, otherwise 0)
    STMNT
    Index of superior (i.e. including or calling) statement in the statement table (>= 1, if DEPTH >= 2, otherwise 0)
    FROM
    Index of first statement of source code unit in the statement table (>= 1)
    TO
    Index of last statement of source code unit in the statement table (>= 1)
    If the source code unit contains include programs or macro calls, the line range [ FROM, TO] in the statement table also covers the statements in subordinate source code units.
    <b>Addition 5</b>
    ...  STRUCTURES INTO itab6
    Details of the construction of the source text table are given in the structure table itab6.
    The structure table itab6 must have the structure SSTRUC.
    The fields in SSTRUC (which are also the columns of structure table itab6) have the following meanings:
    TYPE
    Type of the structure with possible values:
    P (Beginning of the source code)
    R (Subroutine)
    M (Macro, EXEC SQL)
    I (Loop)
    A (Case distinction)
    C (Condition in a case distinction)
    J (Goto command)
    D (Structured declaration)
    E (Event)
    S (Follow-on from simple structured statement)
    STMNT_TYPE
    The statement type of the beginning of the structure. The values are listed in the type pool SCAN in structure SCAN_STRUC_STMNT_TYPE.
    KEY_START
    Flags whether the start of the structure is described semantically ('X' if there is a special statement, otherwise ' ').
    KEY_END
    Flags whether the end of the structure is described semantically ('X' if there is a special statement, otherwise blank).
    STMNT_FROM
    Index of the first statement of the structure in the statement table itab3.
    STMNT_TO
    Index of the last statement of the structure in the statement table itab3.
    Index of the first substructure of the structure in structure table itab6.
    STRUC_TO
    Index of the last substructure of the structure in structure table itab6.
    BACK
    Index of the structure in the structure table itab6 that contains the structure as a substructure (0 if the structure is the root structure of a structure tree).
    <b>Addition 6</b>
    ... OVERFLOW INTO c1
    The addition is only allowed and required if the token table itab2 has the structure STOKEN or STOKEX.
    If a token is too large to be stored in the token table in the field STR, it is placed in the overflow area c1. The offset of the token in the overflow area then lies in the token table in the field OFF1.
    <b>Addition 7</b>
    ... WITH ANALYSIS
    Breaks down each token t = a+b(c) according to the logic of the RSYN key word >ANALY into its three components a, b and c.
    Offset and length of components a, b and c are stored in the fields LEN1, OFF2, LEN2, OFF3, and LEN3 in the token table. (The offset of OFF1 is always 0 and therefore not required.)
    If you specify the addition WITH ANALYSIS, the token table itab2 must have the structure STOKESX, so that the fields LEN1, OFF2, LEN2, OFF3 and LEN3 are available.
    If the token table has the structure STOKEX, you must consider the following:
    If the whole token exists in the token table, the offset specifications are relative to the token start. If the token is in the overflow area c1, the offset specifications are relative to the start of the overflow area.
    <b>Addition 8</b>
    ... WITH COMMENTS
    Returns comments also, with each individual comment representing a token. The system additionally stores entries for each full block of comments in the table itab3, differentiating between comments that occur within statements and those that occur at program level. In itab3, an entry for a comment within a statement always comes before the statement containing the comment.
    <b>Example</b>
    Look at the following program fragment. The preceding numbers are the indexes of the tokens.
    1    * An example  *
    2    * with scattered comments
    6    MOVE
    3    * Inserted comment 1
    7    X
    4    *  Inserted comment 2
    8    TO
    9    Y
    5    * Inserted comment 3
    SCAN then enters the following values for the components TYPE, FROM and TO (in this order from left to right) into itab3.
    'P' 1 2
      'S' 3 5
      'K' 6 9
    If the addition ... WITH COMMENTS is used, the table itab2 must have the line type STOKES or STOKESX.
    <b>Addition 9</b>
    ... WITH INCLUDES
    Also breaks down subordinate source code units (included programs, called macros) into tokens.
    You should normally combine the addition WITH INCLUDES with the addition LEVELS INTO itab5.
    If (at least) one included program does not exist, SY-SUBRC is set to 1 and the relevant INCLUDE statement is flagged in the statement table itab3 by the statement type J (instead of I), but the breakdown process continues. The level table itab5 contains no entry for include-programs that do not exist.
    If you combine WITH INCLUDES with WITHOUT TRMAC , TRMAC-Macros are not expanded because the system does not recognize them as subordinate source code units.
    When macro calls are expanded, no position specifications are available. The corresponding fields in the token table itab2 and the statement table itab3 are then set to 0.
    <b>Addition 10</b>
    ... WITH TYPE-POOLS
    This addition has the same effect as the WITH INCLUDES addition, except that with the former include programs belonging to type groups are broken down into tokens.
    <b>Addition 11</b>
    .. WITH LIST TOKENIZATION
    Tokens of the form (a1, a2, a3) are not returned as tokens but broken down into the elementary components.
    <b>Addition 12</b>
    ... WITHOUT TRMAC
    If a statement begins neither with an ABAP/4 key word nor with a DEFINE macro, the system does not check whether this is a TRMAC macro, but assumes an unknown statement. (Unknown statements are flagged in the statement table itab3 with a U in the field TYPE.)
    To avoid unnecessary database accesses to the table TRMAC, you should use the addition WITHOUT TRMAC whenever you assume that the source code to be scanned contains unknown statements. Unknown statements are particularly likely to occur if you use the addition FROM n1, because the scanner does not start at the beginning of the source code, but from a specified point.
    If you use WITHOUT TRMAC with WITH INCLUDES, TRMAC macros are not expanded because the system does not recognize them as subordinate source code units.
    <b>Addition 13</b>
    ... PROGRAM FROM c2
    <b>Addition 14</b>
    ... INCLUDE INTO c3
    <b>Addition 15</b>
    ... MESSAGE INTO c4
    <b>Addition 16</b>
    ... WORD    INTO c5
    <b>Addition 17</b>
    ... LINE    INTO n3
    <b>Addition 18</b>
    ... OFFSET  INTO n4
    The above additions have the same meaning as those for the
    SYNTAX-CHECK: statement: c2 is an input field for a program name to be assigned to the source code, while the fields c3, c4, c5, n3 and n4 are output fields in case an error occurs.
    To be able to analyze errors without modifying programs, use the additions INCLUDE, MESSAGE, WORD, LINE and OFFSET. These provide information about the errors which have occurred.
    <b>Variant 2</b>
    SCAN AND CHECK ABAP-SOURCE itab1 ...RESULT INTO itab2.
    Parts marked with " ..." are interchangeable
    <b>Extras:</b>
    1. ... PROGRAM FROM c1 2. ... INCLUDE INTO c2
    3. ... MESSAGE INTO c3
    4. ... WORD    INTO c4
    5. ... LINE    INTO n1
    6. ... OFFSET  INTO n2
    The syntax of the program in table itab1 is checked. During the check, all of the information from the program, such as statement structures, statements, tokens, data objects, types and do on are placed into the result field. This field must have the type SYSCH_RESULT, which is defined in type group SYSCH. You must therefore declare type group SYSCH in your ABAP-program using a TYPE-POOLS statement.
    &ABAP_ADDITION _1&
    ... PROGRAM FROM c1
    &ABAP_ADDITION _2&
    ... INCLUDE INTO c1
    &ABAP_ADDITION _3&
    ... MESSAGE INTO c3
    &ABAP_ADDITION _4&
    ... WORD    INTO c4
    &ABAP_ADDITION _5&
    ... LINE    INTO n1
    &ABAP_ADDITION _6&
    ... OFFSET  INTO n2
    The above additions have the same effect as the corresponding additions in the statement SYNTAX-CHECK: c1 is an input field for a program name to be assigned to the source code, the fields c2, c3, c4, n1 and n2 are output fields, used when errors occur.
    To enable you to analyze errors without having to modify the program, you should specify the INCLUDE, MESSAGE, WORD, LINE and OFFSET additions for the information about the error that occurred.
    <b>Reward if usefull</b>

  • What is the use of Node Manager?

    OBIEE Oracle Business Intelligence 11.1.1.5
    I have started up the WebLogic server which delivers the WL Server Admin Console as well as the Fusion Middleware Control.
    I have ignored NodeManager. It hasnt been started.
    What is the use of NodeManager?
    Oracle tells us:
    Note: In order to start the Managed Server using the WebLogic Server Administration Console, the Node Manager must be running.
    http://download.oracle.com/docs/cd/E21043_01/bi.1111/e10541/components.htm
    But my Manged Server is running ok - I can check http://<host>:7001/console and its ok.
    So why do I need to bother with NodeManager?
    I only have WLS and OHS installed - no BI instance.
    DA

    Answers inline:
    As I have installed weblogic server 10.3.5 on windows. while am starting it using cmd am getting two ips as
    1**.*.... for weblogic server one is (channel default 4... .. and one is default 1) kindly explain.
    - It doesn't generate any IP. If you have installed successfully, weblogic will run on your system and it's ip will remain same. You can access it using your systems'IP:portnumber
    Also tell as I have to use cmd for starting servers how can i use unix commands for the same??
    - You have installed on windows so unix command won't work for you. You can either start by going into services window in windows or start using command by going to path weblogichome\domains\domainname\bin. Execute startWebLogic.cmd
    Also explain what is the use of node manager and weblogic scripting tool which are also installed in the system.
    - Node manager is used for communication bentween two or more servers in a domain.
    As am looking to install OIM 11g so kindly tell what are all the things I need to know about weblogic so that i can learn OIM easily.
    - To install OIM, install weblogic, install oracle database. Use RCU utilities to create database and then install SOA, OIM and configure them. You can get all kinds of guides and steps for OIM 11g r2 here:
    http://docs.oracle.com/cd/E27559_01/index.htm
    regards,
    GP

  • What is the use  implemention of work flow

    Hi guis,
      can u plz tell me what is the uses, u implement a workflow, and also please give the one scenrio on purchase order or invoice or anything, one scenrio explain me.

    Workflow is a tool for automating the business process so that the process itself flows smoothly, without bottlenecks and without dead ends. Critical parts of the process can be streamlined with deadline monitoring and contingency options. Participants are informed of milestones in the process. The process definition itself is powerful enough to allow business re-engineering with re-training and to eliminate the impact of workforce changes on the basic business process.
    SAP Business Workflow is the R/3 tool for handling the process automation within R/3 or between R/3 systems and other systems involved in the business process. It is flexible enough to handle ad hoc processes such as the circulation of an electronic document. It is robust enough to handle mission critical process in the production system such as the creation of a material master.
    The different R/3 applications supply standard workflows for the commonly occurring processes. Once these workflow templates are activated, they are ready for immediate use. A complete set of workflow tools, including the workflow editor and workflow generation wizard, are provided to enhance these standard workflows or create new workflows. These tools are complemented with transactions for monitoring, tracking and the statistical analysis of the processes.
    Because the business process often does not step at the R/3 boundary, various interfaces and third party tools are available for extending the process across systems.
    Workflow is the technique for automating business processes. It is particularly beneficial
    u2022 when different people are involved in different parts of the process
    u2022 when a process only occurs sporadically and the users need to be reminded of what they need to do
    u2022 when the duration of the process is critical and deadlines have to be met
    u2022 when the process definition is likely to change on the fly and retraining users is not possible.
    By automating the process, you free the users from having to know what to do when. Work items are sent to the participants notifying them of what they have to do, and a single click on the notification calls the correct transaction and navigates to the relevant screen.
    When the workflow runs it creates work items which are like e-mails, but
    u2022 they have the advantage that they contain the intelligence needed to execute the relevant task with the correct data when the operational user executes them,
    u2022 they provide their own functionality so that the operational user has access to everything that is useful to improve the flow of the process,
    u2022 they disappear on their own when they have been executed by someone else or made obsolete.
    These work items can be received in the R/3 business workplace, MS Outlook (or any other MAPI compatible client), Lotus Notes, or work item notifications can be distributed by e-mail.
    SAP Business Workflow:
    u2022 accelerates the workflow across applications, functional areas, organizational units, and time periods,
    u2022 supplements the integrated management of processes in the SAP System with individually tailored options for automating and actively controlling business and communication processes,
    u2022 combines integrated organizational management with flexible assignment of tasks and responsibilities to locations, positions, jobs, groups, or individuals,
    u2022 actively supports a quick implementation of workflow technologies through predefined workflow templates, Workflow Wizards and flexible adjustment options,
    u2022 integrates every employee in the value chain by providing a versatile integrated inbox ("universal inbox") at their work centers that can be configured individually,
    u2022 provides a Workflow Information System that is the foundation for process change management and continuous business engineering,
    u2022 was designed to be a change-management tool,
    u2022 simplifies continuous business engineering and process change management with user-friendly tools for defining and dynamically changing workflow tasks
    u2022 conforms to the workflow standards of the Workflow Management Coalition (WfMC)
    u2022 interfaces to other mail clients such as Lotus Notes and MS Outlook
    Check these links.
    http://www.sapgenie.com/workflow/index.htm
    Link workflow business objects to your collaboration tasks!!
    http://help.sap.com/saphelp_nw04/helpdata/en/92/bc26a6ec2b11d2b4b5006094b9ea0d/content.htm
    http://help.sap.com/saphelp_bw33/helpdata/en/92/bc26a6ec2b11d2b4b5006094b9ea0d/content.htm
    http://help.sap.com/saphelp_bw31/helpdata/en/8d/25f94b454311d189430000e829fbbd/content.htm
    http://www.sap-press.com/product.cfm?account=&product=H950
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/PSWFL/PSWFL.pdf
    http://www.workflowing.com/id18.htm
    http://www.e-workflow.org/
    http://www.sap-img.com/workflow/sap-workflow.htm
    http://help.sap.com/saphelp_47x200/helpdata/en/a5/172437130e0d09e10000009b38f839/frameset.htm
    http://www.sap-img.com/workflow/sap-workflow.htm
    http://help.sap.com/saphelp_47x200/helpdata/en/a5/172437130e0d09e10000009b38f839/frameset.htm
    Hope this resolves your query.
    raj

  • WHAT IS  THE USE OF COPY/MATCH OPTION  IN BUISNESS CONTENT

    WHAT IS  THE USE OF COPY/MATCH OPTION  IN BUISNESS CONTENT  GIVE DETAILS

    Hi,
    Match (X) or Copy
    If the SAP delivery version and the active version can be matched, a checkbox is displayed in this column.
    With the most important object types, the active version and the SAP delivery version can be matched.
    From a technical point of view, the SAP delivery version (D version) is matched with the M version. As in most cases the M version is identical to the active version (A version) in a customer system, this is referred to as a match between the D and A versions for reasons of simplification.
    When a match is performed, particular properties of the object are compared in the A version and the D version. First it has to be decided whether these properties can be matched automatically or whether this has to be done manually. A match can be performed automatically for properties if you can be sure that the object is to be used in the same way as before it was transferred from Business Content.  When performing matches manually you have to decide whether the characteristics of a property from the active version are to be retained, or whether the characteristics are to be transferred from the delivery version.
    Example of an automatic match
    Additional customer-specific attributes have been added to an InfoObject in the A version. In the D version, two additional attributes have been delivered by SAP that do not contain the customer-specific attributes.  In order to be able to use the additional attributes, the delivery version has to be installed from Business Content again. At the same time, the customer-specific attributes are to be retained. In this case, you have to set the indicator (X) in the checkbox.  After installing the Business Content, the additional attributes are available and the customer-specific enhancements have been retained automatically. However, if you have not checked the match field, the customer-specific enhancements in the A version are lost.
    Example of a manual match
    An InfoObject has a different text in the A version than in the D version. In this case the two versions have to be matched manually. When Business Content is installed, a details screen appears which asks you to specify whether the text should be transferred from the active version or from the D version.
    The Match indicator is set as default in order to prevent the customer version being unintentionally overwritten. If the Content of the SAP delivery version is to be matched to the active version, you have to set the Install indicator separately.
    The active version is overwritten with the delivery version if
    ¡        the match indicator is not set and
    ¡        the install indicator is set.
    In other words, the delivery version is copied to the active version.
    If the Install indicator is not set, the object is not copied or matched. In this case, the Match indicator has no effect.
    In the context menu, two options are available:
                                a.      Merge All Below
    The object in the selected hierarchy level and all objects in the lower levels of the hierarchy are selected as to Match.
                                b.      Copy All Below
    The Match indicators are removed for the object in the selected hierarchy level and all objects in the lower levels of the hierarchy. If the Install indicator is also set, these objects are copied from the delivery version to the active version.
    The most important properties which are taken into account when versions are matched are now listed.
    When referring to InfoObjects it is important to differentiate between:
    §         Characteristics
    §         Time characteristics
    §         Key figures and
    §         Units
    Valid for All InfoObjects Named Above:
    Properties Matched Automatically:
    Properties Matched Manually:
    ·        Texts
    ·        Description
    If texts/descriptions in the delivery version do not yet exist in the active version, they are transferred.
    ·        Attribute only (flag)
    ·        Short and long text (description)
    ·        Time dependency of attributes
    ·        Transfer routine
    Valid for Characteristics
    Properties Matched Automatically:
    Properties Matched Manually:
    ·        Compounding
    New InfoObjects are included in compounding
    ·        Referenced characteristic
    Reference characteristics are transferred automatically. A warning is issued.
    ·        Attributes / navigation attributes
    New attributes are included.
    If an attribute is added by the customer, the settings for this attribute are transferred (for example, F4 sequence).
    ·        Application components
    If the entry for the active version is initial, the delivery version is used.
    ·        BEx display
    ·        Characteristic constants
    ·        Default member
    ·        Authorizations (flag and field)
    ·        Selection (CHASEL)
    ·        InfoObject is document attribute
    ·        Data type
    ·        Length
    ·        Output length
    ·        Conversion routines
    ·        Lower case allowed (flag)
    ·        Contains master data (flag)
    ·        Master data is authorization relevant (flag)
    ·        Export DataSource (Flag)
    ·        F4 query design
    ·        F4 query runtime
    ·        Check ODS object
    ·        Contains text tables (flag)
    ·        Texts are language-dependent (flag)
    ·        Short, medium and long texts exist (each with flag)
    ·        Default currency
    ·        GIS attribute
    ·        Hierarchies exist (flag)
    Valid for Key Figures
    Properties Matched Automatically:
    Properties Matched Manually:
    ·        Type of key figure
    ·        Fixed unit
    ·        Fixed currency
    ·        Aggregation
    ·        Exception aggregation
    Valid for Units
    Properties Matched Automatically:
    Properties Matched Manually:
    ·        See the table above (Valid for Characteristics), if these properties are also relevant for the units.
    When referring to InfoProviders it is important to differentiate between:
    ¡        MultiProviders
    ¡        InfoCubes
    ¡        ODS Objects and
    ¡        InfoObjects
    InfoObjects are discussed above (see the first table Valid for All Objects Named Above).
    Valid for All InfoProviders Named Above:
    Properties Matched Automatically:
    Properties Matched Manually:
    ·        New InfoObjects
    If InfoObjects in the delivery version do not yet exist in the active version, they are transferred.
    ·        Manual matching is not necessary.
    Valid for MultiProviders
    Properties Matched Automatically:
    Properties Matched Manually:
    ·        New InfoProviders
    If InfoProviders in the delivery version do not yet exist in the active version, they are transferred.
    ·        Manual matching is not necessary.
    Valid for InfoCubes
    Properties Matched Automatically:
    Properties Matched Manually:
    ·        New dimensions
    If dimensions in the delivery version do not yet exist in the active version, they are transferred.
    ·        Manual matching is not necessary.
    With the InfoObject catalog object type, InfoObjects that are assigned to a catalog in the delivery version are transferred to the active version automatically. Manual matching is not necessary.
    With the transfer rule object type, those objects that exist in the delivery version but are not yet contained in the active version (for example, additional attributes) are added automatically when the match is performed.  Manual matching is not necessary.
    With the process chain object type, those objects that exist in the delivery version but are not yet contained in the active version (for example, additional objects in the process chain) are added automatically when the match is performed.  Manual matching is not necessary.
    ·        Active Version Available
    In this column, the following display options are available:
    ¡        : The object is available in an active version. You decide whether you want to retain this version or reinstall the latest version of the object.
    ¡        Date symbol: The object is already available in an active version. However, the active version belongs to an older Content release. SAP recommends that you install the latest version.
    ¡        Gray symbol: An activated version of the object is available in the system. However, the object status is inactive. If an object version is activated, this indicates that all metadata for the object can be created successfully. The inactive object status indicates that the processed object is inconsistent and cannot currently be used for processing data. Transfer the object again from the SAP delivery version (D version).  If this is not enough to change the object status from “inactive” to “active” the object has to be postprocessed. You find notes on postprocessing in the activation log.
    Postprocessing: An inactive object status may also occur for an InfoObject if changes are made to the structure at field level. Postprocessing the object involves converting the relevant tables in accordance with the activation log.
    ¡        No indicator: The object is not available in an active version.
    7. Make Settings in the Selection List and Install.
    Make the required settings in the  Install selection list:
    Installation Type
    Installation Type
    Information
    Simulate installation
    The system runs a test to see whether any errors are likely to occur during the installation. However, not all errors that may occur can be identified during simulation: Some error messages can only be generated when a real A version (and not just a simulation) is available in the system. You should, therefore, only use the simulation function as a rough guide to help you identify and remove basic errors.
    Install
    The selected objects are installed immediately.
    Install in background
    The selected objects are installed in the background.
    Install and transport
    The selected objects are installed immediately and then written to a transport request.
    SAP recommends you use the Install in background option because installing a large number of objects can take a long time. Modifying objects that are already active can take an especially long time to complete. Make sure that when you install objects in the background that the versions are not matched. It is always the delivery versions that are installed. These are not matched with any available customer versions.
    Refer
    Business Content (Versions)
    http://help.sap.com/saphelp_nw04/helpdata/en/c1/ea683cc5e8ca68e10000000a114084/frameset.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/80/1a66d5e07211d2acb80000e829fbfe/frameset.htm
    Thanks,
    JituK

  • What is the use of BI ACCELERATOR

    What is the use of BI ACCELERATOR

    Hi,
    B.I Accelerator
    BI accelerator is a separate hardware which when attached to BW server creates an index on data of infocube and stores in compressed format.Different technologies are used for compression and horizontal pratitioning of data.
    BI Accelerator can benefit businesses that have high volumes of data. The read
    performance of BI queries is greatly improved using BI Accelerator. This new
    tool enables quick access to any data with a low amount of administrative effort
    and is especially useful for sophisticated scenarios with unpredictable query types,
    high data volume and high frequency of queries.
    BI Accelerator is also useful when aggregates or database indexes are not
    sufficient, or when these methods become too complex or too costly to maintain.
    For example, it may not be practical or cost effective to create and maintain all the
    aggregates that are needed to support performance for all queries for an InfoCube.
    BI Accelerator Pre requisites
    -->BI Accelerator must run on a Suse Linux Enterprise Server 9.
    -->An RFC Destination must be created between the BI System and BIA.
    -->BI InfoCubes that contain non-cumulative key figures are not supported by BI Accelerator at this time.
    -->BI Accelerator requires SAP Netweaver 2004s SP4 or higher.
    There are two major components in the BI Accelerator architecture that contribute
    to the high performance:
    . BI Accelerator Engine
    . BI Accelerator Indexes
    The BI accelerator is fast enough to aggregate BIA indexes during run time.
    A BI Accelerator index contains lists of Index names, join conditions, join paths, view attributes, semantic relations, and key figures.
    BI Accelerator indexes are independent indexes that store meta data, as well as, execute queries.
    BI Accelerator is a very robust and stable environment that uses simple algorithms versus volatile DB optimizations.
    BI Accelerator is also model-driven:
    integrates NetWeaver BI with the TREX search engine, BIA indexes are a special case of the ESA business object, distinctive SAP technology is leveraged.
    TREX and BI Accelerator
    TREX : Search and classification engine, mainly for unstructured data
    BI Accelerator is built using TREX technology (indexing, retrieving, compression ...) to handle structured data
    Note: BI Accelerator and TREX are two different installations.
    The BI Accelerator cannot be used for standard KMC functionality 
    A TREX installation cannot be used as a BI accelerator
    Technically there is a lot of commonality but TREX-BI Accelerator role switching is not supported.
    BI Accelerator index: Structure containing replicated data and indexes (one index per table) that
    represents the BI star schema for the BI Accelerator
    Indexing: Transfer of InfoCube data to the BI Accelerator server and
    processing and compressing it into the BI Accelerator index.
    To use a BI accelerator index for an InfoCube when you execute a query, you first
    have to activate it and then fill it with data The system performs the following steps
    in order to create an index
    Create
    Index
    Prepare optimize
    Commit optimize
    Advantage Of B.I Accelerator
    Advantage of this is that whenever a query is executed ,Data is fetched from BI accelerator instead of reading of reading frm database reads.
    Thus DB read time is 0 in total query run time leading to improved query performance and stable query response time.
    Increased end-user satisfaction with consistently high query performance.
    Significant reduction in the total cost of ownership: low maintenance, high scalability.
    The performance gain using BI Accelerator can be very great, boosting the response time by a factor of 10-100
    Check this presentation for more details:
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/3604c604-0901-0010-f0aa-b37378495537
    SAP Business Intelligence Accelerator : A High - Performance Analytic Engine for SAP Ne tWeaver Business Intelligence
    http://www.sap.com/platform/netweaver/pdf/BWP_AR_IDC_BI_Accelerator.pdf
    https://www.sdn.sap.com/irj/sdn/bia
    thanks,
    JituK

Maybe you are looking for

  • Trouble with ipod updater

    My ipod hasnt been recognized by itunes so i've been trying the 5 r's. I've tried all but restoring my ipod and none of them have worked. So i installed the updater just now and it doesnt work. It says there ''Ipod service erroe''. Does anyone now ho

  • Why do multiple PXI modules in a chassis cause my PC to continuall​y reset?

    I have a PXI-1000 connected to my desktop PC via a MXI-3 interface. When I plug more than a couple of PXI modules into the chassis and switch on, my PC continually resets itself. Any idea what is wrong?

  • Twitter Widget

    Can anyone save their credentials on the twitter widget? It used to save before the twitter widget was updated. Also, parental controls for the twitter widget seems to have become optional, contrary to the suggestion made in the widget that it is req

  • IBooks has been removed from iPad 1.

    I have the original ipad, and I've just done an update and it's lost iBooks as the new version has been updated to runs on iOS 7 and mine is iOS 5.1.1. Over time I'm losing loads of apps as they upgrade to iOS7, but the loss of iBooks to me is very i

  • Help! I cancelled migration and can´t find the files transferred

    Hi, I just got a new Macbook Pro retina and tried migrating with my old Macbook Pro 2010. I then cancelled the migration. After cancellation, I noticed that the Space Utilized on the disk is 118.83GB, which means that some files were transferred befo