Difference between bexbrowser and Bex analyzer

Hi,
   can any body tell me  what is the difference between bex browser and bex analyzer  and how end users will access the reports and how they access SAP.

Hi
*BEx Web Analyzer *
The BEx Web Analyzer is a standalone, convenient Web application for data analysis that you can call using a URL or as an iView in the portal.
The Web Analyzer allows you to execute ad hoc analyses on the Web: When you have selected a data provider (query, query view, InfoProvider, external data source), the data is displayed in a table with a navigation pane. You can navigate to the data and use other Web Analyzer functions available in the application toolbar. For example, you can change the type of data display, use the information broadcasting functions to broadcast your analyses to others, and create printable versions of your analyses.
In the Web Analyzer, you can save the data view generated from navigation and analysis as a query view by choosing Save View in the context menu, and you can save the ad hoc analysis by choosing Save As. When the query view is saved, only the data view is saved; when the ad hoc analysis is saved, the entire Web application is saved, including the properties of Web items and the layout of the data.
Check the link for more info
http://help.sap.com/erp2005_ehp_03/helpdata/EN/00/e8d13f7fb44c21e10000000a1550b0/frameset.htm
Bex Browser
The Business Explorer Browser (BEx Browser) makes it possible for you to access all document types of the Business Information Warehouse that are assigned to your role or that you have stored in your favorites. You can select and open documents assigned to you in the BEx Browser or store and manage new documents in the BEx Browser.
Document types that you can work with in the BEx Browser are:
·        BW workbooks
·        Documents that are stored in the Business Document Service (BDS)
·        Links (references to file system, shortcuts)
·        Links to internet sites (URLs)
·        SAP transaction calls.
·        Web applications and Web templates
·        Crystal Reports
Regards
Shilpa

Similar Messages

  • What is the difference between upgradation and migration.

    Hi Guru's
    what is the difference between upgradation and migration.
    actuallly i involved in upgradation project, here my role is
    1. first i check the query's in 3.5 save the query and transport the query. and check the query in bex analyzer also.
    2. go to BI .7  find the query;s ,give the query name and save the query ,
    3. once save the query, again will come to 3.5 open the query , it will not open. this is my job here,
        come to 7.0 check the query in analyzer also.
    i am having littile bit confusion, how it will comes query in 7.0, why are u saving the query's in 3.5 and 7.0
    query's already available in 7.0 why are u doing this work?
    can i know the upgrades those  objects, is it neccessary, if necessary how can i upgrade.
    infoobje , transferrules, transferstructure ,infosoure, datasoure,updaterules, ods, cubes.
    Points will be Assingned ,
    Thanks & Regards
    prabhavathi

    Hi,
    I was talking in a general sense not on a query level.
    If your taling about migration in that level meaning as a part of larger upgradation (in your case 3.x to 7) there may be many places where you need to do this kind of activities.
    Fr eg migration into new data flow, Migration of Web templates from BW 3.x to Netweaver 2004s, etc
    Hope this helps.
    Thanks,
    JituK

  • What is the difference between DSo and Infocube

    Hello,
             Kindly tell me what is the difference between DSO and Infocube?
    And please tell me how to take the desicion that in whichi case we can use DSO and in which case we can use Infocube..

    Hi ,
    DataStore object serves as a storage location for consolidated and cleansed transaction data or master data on a document (atomic) level.
    This data can be evaluated using a BEx query.
    A DataStore object contains key fields (for example, document number/item) and data fields that can also contain character fields (for example, order status, customer) as key figures. The data from a DataStore object can be updated with a delta update into InfoCubes and/or other DataStore objects or master data tables (attributes or texts) in the same system or across different systems.
    Unlike multidimensional data storage using InfoCubes, the data in DataStore objects is stored in transparent, flat database tables. The system does not create fact tables or dimension tables.
    Use
    The cumulative update of key figures is supported for DataStore objects, just as it is with InfoCubes, but with DataStore objects it is also possible to overwrite data fields. This is particularly important with document-related structures. If documents are changed in the source system, these changes include both numeric fields, such as the order quantity, and non-numeric fields, such as the ship-to party, status and delivery date. To reproduce these changes in the DataStore objects in the BI system, you have to overwrite the relevant fields in the DataStore objects and set them to the current value. Furthermore, you can use an overwrite and the existing change log to render a source delta enabled. This means that the delta that is further updated to the InfoCubes, for example, is calculated from two successive after-images.
    An InfoCube describes (from an analysis point of view) a self-contained dataset, for example, for a business-orientated area. You analyze this dataset in a BEx query.
    An InfoCube is a set of relational tables arranged according to the star schema: A large fact table in the middle surrounded by several dimension tables.
    Use
    InfoCubes are filled with data from one or more InfoSources or other InfoProviders. They are available as InfoProviders for analysis and reporting purposes.
    Structure
    The data is stored physically in an InfoCube. It consists of a number of InfoObjects that are filled with data from staging. It has the structure of a star schema.
    The real-time characteristic can be assigned to an InfoCube. Real-time InfoCubes are used differently to standard InfoCubes.
    ODS versus Info-cubes in a typical project scenario
    ODS
    why we use ods?
    why is psa  & ods nessasary
    Hope this helps,
    Regards,
    CSM Reddy

  • Office 2013 installed with KB2817430, windowsupdate still finds KB2850036. difference between KB2817430 and KB2850036

    Hi,
    I've made a silent Office 2013 installer with /admin and put SP1 MSP's in the updates folder.
    After deployment under installed updates it shows: Service Pack 1 for Microsoft Office 2013 (KB2817430) 32-Bit Edition
    But when I check for new updates, "ServicePack 1 for Microsoft Office 2013 (KB2850036) 32-Bit Edition" is ready for download and install.
    What is the difference between KB2817430 and KB2850036?
    It seems that I cannot download KB2850036 and also integrate it in the silent installation.
    Another question: Is the updates folder from Office only for servicepacks? or can I put also updates and patches MSP files in here? If yes then i've got a lot of files with the same name and different files sizes. For example After extracting SP1 I've got
    the file "proofsp-en-us.msp" which is 10MB and when extracting another patch also i've got a proof-en-us.msp and is 8MB.
    So which one need to go in the updates folder?
    Can anyone help me out?
    thanks

    Hi,
    As can be seen from below KB articles, KB2817430 only contains Service Pack 1 (SP1) for Office 2013, whereas KB2850036 also contains Service Pack 1 (SP1) updates for other Office 2013 related desktop products. For example, Microsoft Office 2013
    Language Pack, Project 2013, SharePoint Designer 2013, Office Proofing Tools 2013, etc.:
    https://support.microsoft.com/en-us/kb/2850036
    https://support.microsoft.com/en-us/kb/2817430
    So if you have any of these products on your computer, Windows Update might offer related updates to you.
    Regarding the Updates folder issue, you can put any Office software updates in it during the initial installation, such as service packs or hotfixes. A newer patch might make changes to a specific file, you can check the version number of that file, and keep
    the newer one in the folder.
    Here is the reference you will need:
    http://blogs.technet.com/b/office_resource_kit/archive/2011/05/05/using-the-updates-folder-to-deploy-software-updates-with-an-office-2010-installation.aspx
    (Note, it was created against Office 2010, however the steps should hold good for Office 2013)
    Regards,
    Ethan Hua
    TechNet Community Support
    It's recommended to download and install
    Configuration Analyzer Tool (OffCAT), which is developed by Microsoft Support teams. Once the tool is installed, you can run it at any time to scan for hundreds of known issues in Office
    programs.
    Please remember to mark the replies as answers if they help, and unmark the answers if they provide no help. If you have feedback for TechNet Support, contact
    [email protected]

  • What's the Difference Between OLAP and OLTP?

    HI,
    What's the difference between OLAP and OLTP ? and which one is Best?
    -Arun.M.D

    Hi,
       The big difference when designing for OLAP versus OLTP is rooted in the basics of how the tables are going to be used. I'll discuss OLTP versus OLAP in context to the design of dimensional data warehouses. However, keep in mind there are more architectural components that make up a mature, best practices data warehouse than just the dimensional data warehouse.
    Corporate Information Factory, 2nd Edition by W. H. Inmon, Claudia Imhoff, Ryan Sousa
    Building the Data Warehouse, 2nd Edition by W. H. Inmon
    With OLTP, the tables are designed to facilitate fast inserting, updating and deleting rows of information with each logical unit of work. The database design is highly normalized. Usually and at least to 3NF. Each logical unit of work in an online application will have a relatively small scope with regard to the number of tables that are referenced and/or updated. Also the online application itself handles the majority of the work for joining data to facilitate the screen functions. This means the user doesn't have to worry about traversing across large data relationship paths. A heavy dose of lookup/reference tables and much focus on referential integrity between foreign keys. The physical design of the database needs to take into considerations the need for inserting rows when deciding on physical space settings. A good book for getting a solid base understanding of modeling for OLTP is The Data Modeling Handbook: A Best-Practice Approach to Building Quality Data Models by Michael C. Reingruber, William W. Gregory.
    Example: Let's say we have a purchase oder management system. We need to be able to take orders for our customers, and we need to be able to sell many items on each order. We need to capture the store that sold the item, the customer that bought the item (and where we need to ship things and where to bill) and we need to make sure that we pull from the valid store_items to get the correct item number, description and price. Our OLTP data model will contain a CUSTOMER_MASTER, A CUSTOMER_ADDRESS_MASTER, A STORE_MASTER, AN ITEM_MASTER, AN ITEM_PRICE_MASTER, A PURCHASE_ORDER_MASTER AND A PURCHASE_ORDER_LINE_ITEM table. Then we might have a series of M:M relationships for example. An ITEM might have a different price for specific time periods for specific stores.
    With OLAP, the tables are designed to facilitate easy access to information. Today's OLAP tools make the job of developing a query very easy. However, you still want to minimize the extensiveness of the relational model in an OLAP application. Users don't have the wills and means to learn how to work through a complex maze of table relationships. So you'll design your tables with a high degree of denormalization. The most prevalent design scheme for OLAP is the Star-Schema, popularized by Ralph Kimball. The star schema has a FACT table that contains the elements of data that are used arithmatically (counting, summing, averaging, etc.) The FACT Table is surrounded by lookup tables called Dimensions. Each Dimension table provides a reference to those things that you want to analyze by. A good book to understand how to design OLAP solutions is The Data Warehouse Toolkit: Practical Techniques for Building Dimensional Data Warehouses by Ralph Kimball.
    Example: let's say we want to see some key measures about purchases. We want to know how many items and the sales amount that are purchased by what kind of customer across which stores. The FACT table will contain a column for Qty-purchased and Purchase Amount. The DIMENSION tables will include the ITEM_DESC (contains the item_id & Description), the CUSTOMER_TYPE, the STORE (Store_id & store name), and TIME (contains calendar information such as the date, the month_end_date, quarter_end_date, day_of_week, etc).
      Database Fundamentals > Data Warehousing and Business Intelligence with Mike Lampa
    Search Advice from more than 250 TechTarget Experts
    Your question may have already been answered! Browse or search more than 25,000 question and answer pairs from more than 250 TechTarget industry experts.

  • Difference between BAPI and IDocs

    hello all
    what is difference between BAPI and IDocs

    Hi,
    BAPI
    One of the big plusses for BAPIs is that the interface and function are not supposed to change.  This is a big plus when you do upgrades or hot packs because the transaction can change (format, required inputs etc) which means you then need to update the call transaction.
    Some of the BAPIs are better documented and easier to use than others. You usually need to perform the BAPI that actually does the COMMIT after you call your BAPI.
    The Program coding for calling a BAPI is usually cleaner than setting up the screen flow etc for the Call Transaction.
    You don't need to worry about special data circumstances interrupting the normal data flow of the screens and causing errors because of that.
    BAPIs probably have better performance since they don't do the screen flow processing.
    In general if the BAPI exists for the transaction you want to perform and you can figure out how to use it the BAPI is probably the best way to go.
    BAPIs are a subset of the RFC-enabled function modules, especially designed as Application Programming Interface (API) to the SAP business object, or in other words: are function modules officially released by SAP to be called from external programs.
    IDOC
    IDocs are text encoded documents with a rigid structure that are used to exchange data between R/3 and a foreign system. Instead of calling a program in the destination system directly, the data is first packed into an IDoc and then sent to the receiving system, where it is analyzed and properly processed.   Therefore an IDoc data exchange is always an
    asynchronous process. The significant difference between simple RFC-calls and IDoc data exchange is the fact, that every action performed on IDocs are protocolled by R/3 and IDocs can be reprocessed if an error occurred in one of the message steps.
    While IDocs have to be understood as a data exchange protocol, EDI and ALE are typical use cases for IDocs. R/3 uses IDocs for both EDI and ALE to deliver data to the receiving system. ALE is basically the scheduling mechanism that defines when and between which partners and what kind of data will be exchanged on a regular or event triggered basis. Such a set-up is called an ALE-scenario.
    The philosophical difference between EDI and ALE can be pinned as follows: If we send data to an external partner, we generally speak of EDI, while ALE is a mechanism to reliable replicate data between trusting systems to store a redundant copy of the IDoc data.   The difference is made clear, when we think of a purchase order that is sent as an IDoc. If we send the purchase order to a supplier then the supplier will store the purchase order as a sales order. However, if we send the purchase order via ALE to another R/3 system, then the receiving system will store the purchase order also as a purchase order.
    Regards,
    Bhaskar

  • Differences between sapr3 and isu

    Hi All,
    Can anybody explain differences  between isu and sapr3
    Thanks,
    Satya

    Hi Satya ,
    SAP R3 is for the General industry/Organization.SAP after studying and Analyzing the Fortune 500 Companies came up with the Best Process that suits most of the Industries however For few of the Industry Segments SAP came up with its VERTICALS like IS-Retail,IS-Utilities,Etc which will suit the industry requirement.
    Few of the Process and Configurations in IS-Utility will differ from that of the SAP-R3
    Rgds
    SVU123

  • Differences between assets and G/L accts

    Differences between assets and G/L accts in the balance carried forward
    Message no. MQ 555
    Diagnosis
    Differences occurred in the balance carried forward when comparing asset summary records and G/L accounts.
    Procedure
    Using program RFEWSBAL, select the accounts for which differences occurred. To do this, proceed as follows:
    1. Changeover package: Active package
    Phase: ANALYZE
    Origin of change: AS
    Proceed
    1. Analyze the accounts containing errors using note 104567.
    If the document number field contains the entry "W004", this means that although a difference exists in the balance carried forward, this difference was charged off at the key date. Following local currency changeover, the system adjusts the currency differences at the key date only, but not in the balance carried forward.
    If the entry "W004" exists for all affected accounts, the status is set to warning instead of error. If the correction at key date described above proves sufficient, you can continue changing over the local currency without additional adjustments in the previous year.
    However, SAP does recommend that you make corrections in the previous year in line with note 104567 since this ensures that the balance carried forward is also reconciled.

    Hello,
    There 2 t codes ABST and ABST2.
    You have issue if ABST does not match.
    ABST2 is as of today t code. Ideally they will match.
    To fix the issue
    execute AFAR
    then execute AFAB repeat run dep.
    Normally your will asset balance will match with GL

  • What is the difference between rfc and idoc

    hi,
       My name is rambabu.can anyone of u please tell me the main differences between idoc and rfc??
    Thanks in Advance

    hi Ramesh,
    Remote Function Calls (RFC) and data exchange through IDoc
    message documents. RFC makes direct and synchronous calls of a program in
    the remote system. If the caller is an external program it will call an
    RFC-enabled function in R/3 and if the calling program is the R/3 system
    it will call an RFC-function in another R/3-system or it will call a
    non-R/3 program through a gateway-proxy (usually rfcexec.exe). BAPIs are
    a subset of the RFC-enabled function modules, especially designed as
    Application Programming Interface (API) to the SAP business object, or in
    other words: are function modules officially released by SAP to be called
    from external programs.
    IDocs are text encoded documents with a rigid structure that are used to
    exchange data between R/3 and a foreign system. Instead of calling a
    program in the destination system directly, the data is first packed into
    an IDoc and then sent to the receiving system, where it is analyzed and
    properly processed. Therefore an IDoc data exchange is always an
    asynchronous process. The significant difference between simple RFC-calls
    and IDoc data exchange is the fact, that every action performed on IDocs
    are protocolled by R/3 and IDocs can be reprocessed if an error occurred
    in one of the message steps.
    While IDocs have to be understood as a data exchange protocol, EDI and ALE
    are typical use cases for IDocs. R/3 uses IDocs for both EDI and ALE to
    deliver data to the receiving system. ALE is basically the scheduling
    mechanism that defines when and between which partners and what kind of
    data will be exchanged on a regular or event triggered basis. Such a
    set-up is called an ALE-scenario.
    Regards
    Sreeram.G.Reddy

  • Difference between BI and BW 3.5

    10 basic difference between BI and BW 3.5

    Hi the following documentation will be helpful to you,
    Regards,
    SUbha
    Pls grant me points if helpful
    This documentation provides you with a starting point for administration and optimal maintenance of the Enterprise Data Warehousing IT scenario. It contains business-intelligence-specific information for various tasks, lists the relevant tools and functions, and provides links to the more detailed documentation that you need to carry out the tasks.
    The Enterprise Data Warehousing IT scenario supports you in setting up and managing a data warehouse in a company-wide environment. It allows you to integrate, transform, consolidate, clean up, store, and stage data of any origin (SAP and non-SAP sources), or age (historic and up-to-date), to be analyzed and interpreted. It provides flexible structures and layers so that you can react quickly to business developments. It also allows businesses to report in real time and improve the design time and runtime of BI models and processes.
    For more information, see Enterprise Data Warehousing.
    Administration of this IT scenario is part of the Running the Enterprise Data Warehouse scenario variant. Therefore the tasks described below refer to this scenario variant.
    Integration
    Enterprise Data Warehousing provides data for the following IT scenarios:
    IT Scenario
    Description
    Information About IT Scenario Administration
    Enterprise Reporting, Query and Analysis
    This IT scenario shows how business experts can use the reporting and analysis tools available in SAP NetWeaver Business Intelligence (BI) to provide BI information customized for the different categories of user within your company.
    Enterprise Reporting, Query, and Analysis
    Business Planning and Analytical Services
    This IT scenario includes processes that collect data from InfoProviders, queries, or other BI objects, convert them using various methods, and write back new information to BI objects (such as InfoObjects or DataStore objects).
    Business Planning and Analytical Services
    Tasks
    In most cases, running an enterprise data warehouse involves general tasks for managing the Application Server for ABAP (AS ABAP) and the Application Server for Java (AS Java), as well as scenario-specific tasks. For more information, see:
    ●     AS ABAP (Application Server for ABAP)
    ●     AS Java (Application Server for Java)
    ●     Enterprise Data Warehousing: Preparation for EDW Processes
    ●     Enterprise Data Warehousing: Tasks Performed Regularly
    ●     Enterprise Data Warehousing: Tasks on Demand
    ●     Enterprise Data Warehousing: Saving and Recovering Data
    For information about the general administration of BI systems, see BI (Business Intelligence).

  • Differences between APO and R3 sysytem.

    Hi everybody,
                        Can anyone list out the differences between the APO system and the R3 system.A detailed explanation will be helpful.

    Hi,
    APO
    APO is the supply chain tool from SAP.
    BW is the data layer for APO.
    SAP APO is for Demand Planning and is a component of mySAP SCM. It enables dynamic supply chain management and which in turn includes applications for detailed planning, optimization and scheduling, allowing the supply chain to be accurately and globally monitored even beyond enterprise boundaries.
    SAP BI is a reporting tool in SAP. It provides a complete view of the company. The solution is part of SAP NetWeaver and provides the tools needed to make the right decisions, optimize processes, and measure strategic success: data from all enterprise sources are merged effectively and can be comprehensively analyzed, business-critical factors can be monitored using external and internal benchmarks, the most important KPIs can be passed on to all the relevant forces in a company in real time, strategies can be rapidly adapted to changed market conditions, and the results of analyses can be graphically represented in a wide variety of formats, such as grids, graphs, and maps
    Check this:
    What is APO and how it is related to BW
    SAP Advanced Planner and Optimizer is a component of the mySAP Supply Chain Management solution that you use for planning and optimizing supply chain processes at a strategic, tactical, and operational planning level.
    With APO - Simultaneous Material & Capacity Planning is possible, as against R/3 PP
    APO Planning is capable of taking all the constraints into consideration with Planning for Capacities/Scheduling
    APO Resource (workcenter in R/3) does a Capacity availability check before scheduling an activity, if capacity is not available, it finds the immediate next available slot & schedules the activity there. This is not possible with R/3
    http://www.sap.com/solutions/business-suite/scm/pdf/BWP_APO40.pdf
    http://help.sap.com/saphelp_apo/helpdata/en/7e/63fc37004d0a1ee10000009b38f8cf/frameset.htm
    http://help.sap.com/bp_biv133/documentation/BW/I18_Scen_InstallGuide_EN_DE.doc
    http://help.sap.com/saphelp_scm41/helpdata/en/92/dd013872af2946e10000009b38f8cf/content.htm
    http://help.sap.com/saphelp_apo/helpdata/en/7e/63fc37004d0a1ee10000009b38f8cf/frameset.htm
    R/3 OLTP
    Online Transactional data processing.
    Designed to efficiently process high volumes of transactions, instantly recording business events (such as a sales invoice payment) and reflecting changes as they occur.
    Current data
    Short database transactions
    Online update/insert/delete
    Normalization is promoted
    High volume transactions
    Transaction recovery is necessary
    http://en.wikipedia.org/wiki/OLTP
    Refer for OLTP Vs OLAP
    compare to r/3, what are the benefits available for b/w
    R/3 and BW reports
    http://www.geekinterview.com/question_details/209
    http://blogs.netindonesia.net/kiki/archive/2006/02/28/8822.aspx
    What's the Difference Between OLAP and OLTP?
    Thanks,
    JituK

  • Minimum difference between average and peak bitrage in mpeg2?

    is there a technical reason why compressor insists on certain minimum differences between average and peak bitrate when compressing to mpeg2?
    when analyzing other footage, i regularly see material where the difference is much smaller than what compressor will allow me to do.
    for instance; i'm seeing material that is compressed at 4.7 average with a peak at 5.2. compressor insists on setting the peak at 5.5 when the average is 4.7 ...

    Well, compression theory says that, ideally, the max bit rate would be double the average bit rate. But since since life often gets in the way of theoretical perfection, Compressor 2 imposes a minimum 1 Mbps difference. This is partially touched upon in the DVDSP 4 manual. From page 64:
    Since the VBR methods are most effective when the maximum bit rate is about double the basic bit rate, or at least 1 to 3 Mbps higher than the basic bit rate, the Max Bit Rate setting is forced to be higher than the Bit Rate setting.
    Now, while Compressor 2 is coded for a 1 Mbps difference, it doesn't always follow that, in my experience. There are instances when, footage permitting, Compressor uses average bit rates lower than specified as well as never hitting the max bit rate that you've set (but it really does depend on the footage).

  • Difference between validation and susbstittution with tcodes

    hi all can any body tell me about difference between validation and susbstittution with tcodes
    regars
    sravya

    Hi,
    Settting up the FI Validations
    You can use additional validation to supplement the existing SAP logic to fit you into your businness needs.
    For e.g. to allow postings from company code 0001 to business area AA only.
    GGB0 - To maintain the validations
    Validation - Enter a validation name e.g. VBUS
    Applicaion Area - SAP module e.g. FI
    Calling Point - Dependent upon the application area that is selected.
    0001 - Document Header - validate enteries at document header, e.g BKPF table
    0002 - Line Items - check line item entris within a document, e.g. BSEG table
    0003 - Complete Document - check settings for the whole documents.
    e.g. Validation step 001
    Allowed company code '0001' to post to business area AA only. 'FI00' checks that line item is from FI.
    If checks failed, the error message number 14 is displayed to stop the transaction.
    Prerequisite
    <BSEG> $BUKRS = '0001' AND
    <BKPF> $GLVOR = 'FI00'
    Check
    <BSEG> $GSBER = 'AA'
    Message
    Type E No. 14 - Business Area not allowed for company 0001
    Output fields 1 - BSEG - GSBER 2 - BSEG - BUKRS
    3 - 4
    OB28 - Define Validations for Posting
    CC CallPnt Validation Description Activtn Level
    0001 2 VBUS Business Area Validation 1
    Activtn Level - 0 - Not Active 1 - Activated 2 - Activated for all except batch input
    GGB4 - Manage the activation of all validations in the SAP system.
    Validations, Substitutions, and Rules
    Purpose
    With the validations and substitutions software, you can validate and/or substitute data at the time of entry in the FI-SL System and other SAP Systems.
    Implementation Considerations
    You do not need to change the standard system to customize your validations and substitutions. For each installation, you can define exactly which validations and substitutions are necessary to ensure the integrity of your data.
    Features
    Validation
    In the SAP System, almost all input values are validated by a program or against tables or master files. Since some types of validations cannot be standardized, you can use FI-SL's validations program to create validations for your system.
    With validations, you can check values and combinations of values as they are being entered in the SAP environment. Validation rules are stored in the Rule Manager; as data is entered, the Integration Manager validates the data against the validation rules stored in the Rule Manager. Because data is validated before it is posted, only valid information enters the FI-SL System.
    You define validation rules using Boolean logic. When you define a validation rule, the system checks the validation rule to ensure that it is syntactically correct.
    You can also define the result of breaking the validation rule. For instance, you can specify that a message is issued but the user can continue processing. You can also force the user to correct the error before processing can continue.
    For more information, see Validation or What Are Validations?. For more information on using validations, see Creating a Validation. For more information on Boolean Logic, see Introduction to Boolean Logic.
    Substitution
    Substitution rules are stored in the Rule Manager. When data is entered in the system, it is substituted by the Integration Manager. The Integration Manager calls the Rule Manager. Substitution occurs before the data is posted to the FI-SL databases.
    You define substitution rules using Boolean logic. When you define a substitution rule, the system checks the substitution rule to ensure that it is syntactically correct.
    For more information, see Substitution or What Are Substitutions?. For more information on using substitutions, see Creating a Substitution. For more information on Boolean Logic, see Introduction to Boolean Logic.
    Rules
    A rule is a Boolean expression that you can use as a condition, as a check, or in another rule. Rules enable you to refer to frequently used Boolean statements simply by specifying the rule name.
    The FI-SL Integration Manager uses Boolean logic. When you use Boolean logic and syntax, you can:
    Validate data
    Substitute data
    Select ledgers for posting
    Transfer substitution data to a rollup ledger
    Select report data
    The Rule Manager stores the Boolean rules used to analyze data. The Integration Manager calls the Rule Manager and determines whether entered data should be used. If a Boolean statement is true, an action is taken; if a statement is false, no action is taken.
    Boolean logic is used in the following FI-SL modules:
    Validation
    Substitution
    Ledger selection
    Report Writer
    Rollups
    To use Boolean Logic in one of the functions listed, you create statements that are used as formulas in the FI-SL System. For more information on using Boolean Logic, see Introduction to Boolean Logic or Boolean Logic Statements and Rules.

  • What is the difference between tkprof and explainplan

    Hi,
    what is the difference between tkprof and explainplan.

    Execution Plans and the EXPLAIN PLAN Statement
    Before the database server can execute a SQL statement, Oracle must first parse the statement and develop an execution plan. The execution plan is a task list of sorts that decomposes a potentially complex SQL operation into a series of basic data access operations. For example, a query against the dept table might have an execution plan that consists of an index lookup on the deptno index, followed by a table access by ROWID.
    The EXPLAIN PLAN statement allows you to submit a SQL statement to Oracle and have the database prepare the execution plan for the statement without actually executing it. The execution plan is made available to you in the form of rows inserted into a special table called a plan table. You may query the rows in the plan table using ordinary SELECT statements in order to see the steps of the execution plan for the statement you explained. You may keep multiple execution plans in the plan table by assigning each a unique statement_id. Or you may choose to delete the rows from the plan table after you are finished looking at the execution plan. You can also roll back an EXPLAIN PLAN statement in order to remove the execution plan from the plan table.
    The EXPLAIN PLAN statement runs very quickly, even if the statement being explained is a query that might run for hours. This is because the statement is simply parsed and its execution plan saved into the plan table. The actual statement is never executed by EXPLAIN PLAN. Along these same lines, if the statement being explained includes bind variables, the variables never need to actually be bound. The values that would be bound are not relevant since the statement is not actually executed.
    You don’t need any special system privileges in order to use the EXPLAIN PLAN statement. However, you do need to have INSERT privileges on the plan table, and you must have sufficient privileges to execute the statement you are trying to explain. The one difference is that in order to explain a statement that involves views, you must have privileges on all of the tables that make up the view. If you don’t, you’ll get an “ORA-01039: insufficient privileges on underlying objects of the view” error.
    The columns that make up the plan table are as follows:
    Name Null? Type
    STATEMENT_ID VARCHAR2(30)
    TIMESTAMP DATE
    REMARKS VARCHAR2(80)
    OPERATION VARCHAR2(30)
    OPTIONS VARCHAR2(30)
    OBJECT_NODE VARCHAR2(128)
    OBJECT_OWNER VARCHAR2(30)
    OBJECT_NAME VARCHAR2(30)
    OBJECT_INSTANCE NUMBER(38)
    OBJECT_TYPE VARCHAR2(30)
    OPTIMIZER VARCHAR2(255)
    SEARCH_COLUMNS NUMBER
    ID NUMBER(38)
    PARENT_ID NUMBER(38)
    POSITION NUMBER(38)
    COST NUMBER(38)
    CARDINALITY NUMBER(38)
    BYTES NUMBER(38)
    OTHER_TAG VARCHAR2(255)
    PARTITION_START VARCHAR2(255)
    PARTITION_STOP VARCHAR2(255)
    PARTITION_ID NUMBER(38)
    OTHER LONG
    DISTRIBUTION VARCHAR2(30)
    There are other ways to view execution plans besides issuing the EXPLAIN PLAN statement and querying the plan table. SQL*Plus can automatically display an execution plan after each statement is executed. Also, there are many GUI tools available that allow you to click on a SQL statement in the shared pool and view its execution plan. In addition, TKPROF can optionally include execution plans in its reports as well.
    Trace Files and the TKPROF Utility
    TKPROF is a utility that you invoke at the operating system level in order to analyze SQL trace files and generate reports that present the trace information in a readable form. Although the details of how you invoke TKPROF vary from one platform to the next, Oracle Corporation provides TKPROF with all releases of the database and the basic functionality is the same on all platforms.
    The term trace file may be a bit confusing. More recent releases of the database offer a product called Oracle Trace Collection Services. Also, Net8 is capable of generating trace files. SQL trace files are entirely different. SQL trace is a facility that you enable or disable for individual database sessions or for the entire instance as a whole. When SQL trace is enabled for a database session, the Oracle server process handling that session writes detailed information about all database calls and operations to a trace file. Special database events may be set in order to cause Oracle to write even more specific information—such as the values of bind variables—into the trace file.
    SQL trace files are text files that, strictly speaking, are human readable. However, they are extremely verbose, repetitive, and cryptic. For example, if an application opens a cursor and fetches 1000 rows from the cursor one row at a time, there will be over 1000 separate entries in the trace file.
    TKPROF is a program that you invoke at the operating system command prompt in order to reformat the trace file into a format that is much easier to comprehend. Each SQL statement is displayed in the report, along with counts of how many times it was parsed, executed, and fetched. CPU time, elapsed time, logical reads, physical reads, and rows processed are also reported, along with information about recursion level and misses in the library cache. TKPROF can also optionally include the execution plan for each SQL statement in the report, along with counts of how many rows were processed at each step of the execution plan.
    The SQL statements can be listed in a TKPROF report in the order of how much resource they used, if desired. Also, recursive SQL statements issued by the SYS user to manage the data dictionary can be included or excluded, and TKPROF can write SQL statements from the traced session into a spool file.
    How EXPLAIN PLAN and TKPROF Aid in the Application Tuning Process
    EXPLAIN PLAN and TKPROF are valuable tools in the tuning process. Tuning at the application level typically yields the most dramatic results, and these two tools can help with the tuning in many different ways.
    EXPLAIN PLAN and TKPROF allow you to proactively tune an application while it is in development. It is relatively easy to enable SQL trace, run an application in a test environment, run TKPROF on the trace file, and review the output to determine if application or schema changes are called for. EXPLAIN PLAN is handy for evaluating individual SQL statements.
    By reviewing execution plans, you can also validate the scalability of an application. If the database operations are dependent upon full table scans of tables that could grow quite large, then there may be scalability problems ahead. On the other hand, if large tables are accessed via selective indexes, then scalability may not be a problem.
    EXPLAIN PLAN and TKPROF may also be used in an existing production environment in order to zero in on resource intensive operations and get insights into how the code may be optimized. TKPROF can further be used to quantify the resources required by specific database operations or application functions.
    EXPLAIN PLAN is also handy for estimating resource requirements in advance. Suppose you have an ad hoc reporting request against a very large database. Running queries through EXPLAIN PLAN will let you determine in advance if the queries are feasible or if they will be resource intensive and will take unacceptably long to run.

  • What's the difference between trace and log in XI?

    Hi all,
    On both J2EE and ABAP, there are both tracing and logging functionalities. Can anyone explain me the difference between trace and logging?
    Regards,
    Hui

    From SAP Notes
    Log file:
    A log file contains generally intelligible information for system administrators. The information is sorted by categories and is used for system monitoring. Problem sources or critical information about the status of the system are logged in this file. If error messages occur, you can determine the software component that has caused the error using the location. If the log message does not provide enough details to eliminate the problem, you can find more detailed information about the error in the trace file.
    The log file is located in the file system under
    "/usr/sap/[SID]/[instance]/j2ee/cluster/server[N]/log/applications.[n].log" for every N server node.
    Access the file with the log viewer service of the J2EE visual administrator or with the standalone log viewer.
    Trace file:
    A trace file contains detailed information for developers. This information can be very cryptic and extensive. It is sorted by location, which means by software packages in the Java environment, for example, "com.sap.aii.af". The trace file is used to analyze runtime errors. By setting a specific trace level for specific locations, you can analyze the behavior of individual code segments on class and method level. The file should be analyzed by SAP developers or experienced administrators.
    The trace file is located in the file system under
    "/usr/sap/[SID]/[instance]/j2ee/cluster/server[N]/log/defaultTrace.[x].trc" for each N server node.
    Access the file with the log viewer service of the J2EE visual administrator or with the standalone log viewer.
    Regards,
    Prateek

Maybe you are looking for