Infosets, MultiProvider differences?

Hi,
I'm confused about why do we need infosets when multiproviders are there. Are there significant differences where in these two are used in different scenarios?
Thanks,
Prashanti

Hi,
For more info on multiproviders:
http://help.sap.com/saphelp_nw04/helpdata/en/52/1ddc37a3f57a07e10000009b38f889/content.htm
For more info on infosets:
http://help.sap.com/saphelp_nw04/helpdata/en/ed/084e3ce0f9fe3fe10000000a114084/content.htm
Please search before posting a query, as these topics have been discussed frequently....
SDN links where you can find detailed info about this:
difference between Infosets & Multiproviders
https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/2f5aa43f-0c01-0010-a990-9641d3d4eef7
Cheers,
Habeeb

Similar Messages

  • Generic Extraction - query view and infoset query difference

    Hi Guys,
    I want to know the Generic datasource difference between Extraction from view and Extraction from Query .I have read some threads and documents, still i am not getting the exact difference.
    My understanding :
    Views: combination of tables to view the end result , we use views.
    Infoset query also with help of SQ01, 02, 03  : we combine tables to see the result.
    both use primary key to select the data.
    Please help me.
    Thanks ,
    Nithi.

    Hi Nithi ,
    When we build the datasource on the View
    View is basically a table in which you can take data from many table combniation  .
    in view you can specify the table join condtion and selection condtion
    View data gets updated as the data in the source table gets updated
    When we build the datasource on the Infoset
    Data source str gets filled from the infoset query .you need to use SQ01 and Sq02 transaction for this
    in infoset you can define the tables from which you want to read the date .There is also facilty to write ABAP code which will process the data and fill the the data source str using infoset query
    infoset query read the data from infoset and transfer to data structure
    Thanks
    Rajendra

  • Prompt List of Values in WebI - SAP BI Variable - SAP BI Query -BI Infoset

    Hi,
    We have a Multiprovider built on an Infoset. SAP BI Infoset is built using 2 DSOs.
    There is a SAP BI Query on top of this Multiprovider.
    Multiple-Single Values type Optional Variables are created for some of the Objects of this SAP BI Query.
    Universe is created on top of this SAP BI Query, using SAP Integration kit.
    WebI is created on top of this Universe.
    Prompts are created on some of the Universe Objects, for the WebI report.
    We are facing a typical problem.
    When a WebI prompt is created on one particular object, (which also has a Multiple Single Value type Optional Variable in SAP BI Query), its List of Values is showing up only first three values in WebI. The SAP BI Variable in BEX Analyzer shows up all the 8 values for the List of Values, for the same object.
    The Output WebI report has more values for the object than shown up in the List of Values & this is the problem.
    We tried to keep the settings in Info-object, DSO, Infoset, Multiprovider = Values in Master Data (for the List of Values).
    Similar objects and similar variables for the same query / webi show all the List of Values.
    Any guidance on aspects to be checked would be helpful.
    regards,
    Rajesh K Sarin

    Problem :
    If an SAP BI Infoset is used in data modelling, and a restricted key
    figure is created in the SAP BI Query; the resultant WebI Object Prompt
    for the Info-object, only shows up LOVs which have been used for the
    restricted Key Figure. Both the things are not related except for the
    fact that some of the values for the same object are used for the
    restricted key figure.
    Details :
    We have a Multiprovider built on an SAP BI Infoset.
    SAP BI Infoset is built using 2 DSOs.
    There is a SAP BI Query on top of this Multiprovider.
    There is a restricted key figure in the sap bi query. KeyFigure "Number
    of Records" is restricted for 3 of the relevant Info-object "Meter
    Reading Status" values.
    Universe is created on top of this SAP BI Query, using SAP Integration
    kit. WebI is created on top of this Universe.
    Prompt is created on the Universe Object for Meter Reading Status, in
    the WebI report.
    List of Values is showing up only the three values in WebI (which were
    used in the Restricted Key Figure calculation).
    SAP BI Variable in BEX Analyzer for the same Info-object Meter Reading
    Status shows up all the 8 values for the List of Values.
    The Output WebI report has more values for the object than shown up in
    the List of Values & this is the problem.
    We have kept the settings in Info-object, DSO, Infoset,
    Multiprovider, Query = Values in Master Data (for the List of Values).
    When we delete the "Restricted Key Figure" from the SAP BI Query, the
    refreshed Universe and Webi have a complete set of List of Values
    applicable.   
    Steps for Reconstruction    
    1. Create Infoset
    2. Create SAP BI Query on Infoset
    3. Create Restricted Key Figure for Number of Records and few of the
    values for a characteristic info-object "X".
    4. Create BO Universe on SAP BI Query
    5. Create Webi on BO Universe
    6. Create a Prompt on the characteristic info-object "X".
    7. Check List of Values for the characteristic info-object "X" in SAP
    BI Query and WebI.
    8. LOVs in WebI are equal to the values used for restriction in the
    restricted key figure.
    9. Delete the restricted key figure from SAP BI Query.
    10. Refresh Universe
    11. Webi prompt shows all the values for the prompt on "X".

  • Currency Translation in Query on InfoSet

    Hi Experts,
    I need to do a currency conversion from 0LOC_CURRENCY to USD on an Infoset. The Data flow is DSOs --> Infoset --> Multiprovider --> Query. I want to do this Translation on Posting date which is in my DSO, Infoset and Multiprovider. I have created custom Translation Type on Postind date (with Exchange Rate Type = M, Fixed Target Currency = USD, Time Reference as Special Infoobject = 0PSTNG_DATE in RSCUR) in RSCUR but in the Query, I cannot see this custom conversion type available, as in report the infoobject 0PSTNG_DATE is available as '<Infoset Name>__FXX' .
    Is there any paricular setting need to be done on an infoset to incorporate Currency Translation on Query ? Also, if there is any other procedure available for Infoset Currency Translation, please let me know.
    Thanks fin advance or all your help.
    Regards,
    Dibyendu

    Hi,
    As I want to do my currency translation on an Infoset the infoobjects do not appear in traditional format. 0PSTING_DATE has taken a name '<Infoset Name>__FXX' (like ZISCURTR__F24). How can I put this reference in RSCUR 'Time Ref' tab in Infoset option over there ? It gives only a F4 help option (donot allow to write something there) and my infoset is not available in the list. Is there any particular setting that I can do in my infoset to make that available in RSCUR F4 help ?
    Please suggest.
    Regards,
    Dibyendu

  • Dates are not mapped as the Date data type in Universe created on Infoset

    Hi,
    When i try to create the universe on top of Infoset query, the date fields present as the variables in SAP BI Query are not mapped as date data type in Universe, instead they are considered as Characters, hence the prompts related to those variables are coming s List of values instead of Calendar in WebI.
    Can any one pls help me to find the solution on how to map date variables as calendar in Webi Prompts.(For SAP BI Queries created on Infoset)
    Edited by: Nisha Makhija on Jul 20, 2009 5:59 PM

    Hi Ingo,
    Thanks for your response!!
    The I Query is built on Top of BW MultiProvider and the InfoObject is of Type DATS.
    Actually our Modeling on the BI is as follows :
    DSOs> Infoset>Multiprovider->BI query>Universe.
    Since we were not able to get the prompts as Calander so We tried Debugging on different Data targets to find the root cause. Please find our observations as follows:
    When I tested the same date infoobject(of Type DATS) by creating a BI Query on top of DSO, Infocube & Multiprovider ,the Prompts are working fine as a calander in WEBI Report.
    But when the BI Query is built on Infoset data target, In WEBI Report I am getting list of values rather a Calender prompt .
    I tested in the Universe that field is appearing as Character instead of Date.
    Please guide to resolve this issue.
    Thanks,
    Nisha.

  • MDX Query Failed to execute when universe created on MultiProvider Query

    Hi All,
    I am running the WebI report created on the universe which is built on Query.
    The data flow model in our system for this query is:
    Data Store Object> Infoset-> multiProvider--->Query.
    When I drag only on Dimension object in the Webi Report without any Prompts the report works fine.
    But when I add any other Dimension or any Prompt in the WebI Query it gives me following error:
    A database error occured. The database error text is: The MDX query SELECT  { [Measures].[4EGM09GXGU939C3RJH2G9DOQY] }  ON COLUMNS , NON EMPTY CROSSJOIN( [Z_WM_IS01___F1].[LEVEL01].MEMBERS, [Z_WM_IS01___F12].[LEVEL01].MEMBERS ) DIMENSION PROPERTIES [Z_WM_IS01___F12].[2Z_WM_IS01___F12], [Z_WM_IS01___F1].[2Z_WM_IS01___F1] ON ROWS FROM [ZWM_M02/Z_ZWM_M02_Q001]  failed to execute with the error Unknown error. (WIS 10901)
    But if i create the Universe on Query which is built on DSO then WebI is working fine.
    Is there any restrictions on using Query on multiprovider?
    Or do I have to make any configuration at BO server?
    How do I configure logging for WebI and Designer to check the Log records?
    Do I have to change anything related to any MDX Parser in BO Server?
    We are using folowing versions:
    SAP BI 7.0 and BOXIR3.1
    Request to pls help me to solve this problem.

    Thanks Ravi for ur inputs.
    The report is working fine now.
    There was the problem with the Mandatory prompts defined in SAP BI  Query on which universe was developed.
    U should have only those Prompts as mandatory in SAp BI Query which are Universe level prompt, if some mandatory prompts defined as Class level promts in universe then the MDX Query generated based on WEBI report will not Execute.

  • Interview Questions

    Hy Gurus
    I am new to BW.
    Can u please let me know the answers for the following Qs.
    1.     What is a Data warehouse? List Few Properties of a Data warehouse.
    2. What are the major challenges of any Data Warehouse design?
    3. Data loading issues, Reporting, Production support issues?
    4.     Data Modeling, Info cube, ODS obj creation?
    5.     SAP Batch process in BW side?
    6.     How to schedule the Background jobs in ABAP, & BW?
    7.     Variables in Query? User exit variables? How do u create user exit variable? And wt u have used?
    8.     Structures in Query Reporting? What are structures?
    9.     Transportation steps?
    10.     How to compress the Info cube? What happens when you compress the info cube?
    11.     Performance issues in Reporting? How can you improve the performance of the query?
    12.     Where the ABAP routines are used in BW side?
    13.     How to create Primary index & secondary indexes on ODS?
    14.     How to prevent the duplicate records at the data target level?
    15.     Issues regarding double records?
    16.     Info sets and multiprovider differences?
    17.     Issues faced in Delta loading? Why it happened?
    Enhancements to the Data sources?
    18.     Issues in loading the data from Flat files? Delta load issues? Which will u suggest for flat file loading is it delta or full load?
    19.     How to prevent the Replication Errors? What happens when you replicate the Data Sources?
    20.     Process chain steps? Which process type is used to delete the last request when loading the data?
    21.     What is virtual cube? Characteristics? Its significance?
    22.     Diff. methods for Generic Data Sources?
    23.     What is Extract structure? Where it is used?
    24.     Data Modeling – design issues, Tech specifications, Modeling , Reporting, Testing, Transportation
    25.     Extraction on R/3 side steps?LO’s
    26.     How do you setup LIS?
    27.     SAP Batch process? /.where does u setup batch process?
    28.     Reconstruction tab in Info cube? Why it is used?
    29.     Suppose Info cube (A) having 10 records.
                         We want to take the some records say 4-7 from Info cube (A) --&#61664; Info cube (B).
                How does u handle the situation?
    30.     Suppose ODS contains 5 records and the Info cube shows 12 records. How to solve it?
    31.     Landscape
    32.     Tell me typical BW Team & How u work?
    33.     How to Maintain the Master data at Client –Server architecture?
    34.     Query performance problems? How does u improve the performance?
    35.     How does u improve the Info cube design performance?
    36.     How do you improve the Dimension & fact table performance?
    37.     How to push data from PSA to ODS?
    38.     How to postpone daily load?
    39.     The functions of Administrator Workbench are…
    40.     What is RECORD MODE
    41.     What is partition? How to partition the Info cube & PSA. ?..
    42.     How to filter single records when uploading into an ods object?
    43.     How can u connect a new data target to an existing data flow?
    44.     When it is advantageous to create secondary indexes for ODS field?
    45.     Purpose of setup tables?
    46.     What is Delta mechanism?
    47.     Will u create an infosourse for every data souse? How many infosourse can we connect to a Data Sources?
    48.     What is a Star Schema? What is the Differences between a Classic Star Schema & Extended Star schema.
    49.     What is an Attribute? Difference between Display Attribute & Navigational Attribute?
    50.     What is Transfer Rule? List the methods used in Transfer Rules.
    51.     Why we need ODS? /List a few of the technical settings that can be defined when building/modifying an ODSObject?
    52.     What is ODS? What are the three tables associated with an ODS object? What are the two types of ODS.
    53.     Name the two tables that provide detailed information about data sources.
    54.     What are two data transfer methods? Which is the preferred method and why?
    55.     Where will the development take place?&  Who will do the development testing.
    56.     Who will be responsible for long-term support?
    57.     What is a Slowly Changing Dimension?
    58.     What is namespace for SAP BW?
    59.     What are nine decision points of Data warehouse
    60.     How does u install BW Statistics? How you enable monitoring using BW Statistics.
    61.     How do u rate u r self in Modeling, Reporting & Extraction.
    62.     What are the advantages with LO Extraction.
    63.     What are the steps in General to enhance data?
    64.     What are Phases in ASAP Methodology?
    65.     What is the 90 Day rule?
    66.     What is the demo content use?
    67.     What is the use of RSRAJ Transaction?
    68.     What is the use of RSSU53 Transaction?
    69.     Can u repeat the Master Data Source?
    70.     What is diff. b/w DW and BW?
    71.     Will u use idol methods in BW?
    72.     What does the number in the 'Total' column in Transaction RSA7 mean?
    73.       The extract structure was changed when the Delta Queue was empty. Afterwards new delta records were written to the Delta Queue. When loading the delta into the PSA, it shows that some fields were moved. The same result occurs when the contents of the Delta Queue are listed via the detail display. Why are the data displayed differently? What can be done?
    74.     How and where can you control whether a repeat delta is requested?
    75.     Can I use several clients within SAP BW?
    Thanks in Advance
    murali

    Hi
    I have few questions and answer u can check it out which would be more helpful
    Solution
    Questions and answers.
    Question 1:
    What does the number in the 'Total' column in Transaction RSA7 mean?
    Answer:
    The 'Total' column displays the number of LUWs that were written in the delta queue and that have not yet been confirmed. The number includes the LUWs of the last delta request (for repeating a delta request) and the LUWs for the next delta request. An LUW only disappears from the RSA7 display when it has been transferred to the BW System and a new delta request has been received from the BW System.
    Question 2:
    What is an LUW in the delta queue?
    Answer:
    An LUW from the point of view of the delta queue can be an individual document, a group of documents from a collective run or a whole data packet from an application extractor.
    Question 3:
    Why does the number in the 'Total' column, in the overview screen of Transaction RSA7, differ from the number of data records that are displayed when you call up the detail view?
    Answer:
    The number on the overview screen corresponds to the total number of LUWs (see also question 1) that were written to the qRFC queue and that have not yet been confirmed. The detail screen displays the records contained in the LUWs. Both the records belonging to the previous delta request and the records that do not meet the selection conditions of the preceding delta init requests are filtered out. This means that only the records that are ready for the next delta request are displayed on the detail screen. The detail screen of Transaction RSA7 does not take into account a possibly existing customer exit.
    Question 4:
    Why does Transaction RSA7 still display LUWs on the overview screen after successful delta loading?
    Answer:
    Only when a new delta has been requested does the source system learn that the previous delta was successfully loaded into the BW System. The LUWs of the previous delta may then be confirmed (and also deleted). In the meantime, the LUWs must be kept for a possible delta request repetition. In particular, the number on the overview screen does not change if the first delta is loaded into the BW System.
    Question 5:
    Why are selections not taken into account when the delta queue is filled?
    Answer:
    Filtering according to selections takes place when the system reads from the delta queue. This is necessary for performance reasons.
    Question 6:
    Why is there a DataSource with '0' records in RSA7 if delta exists and has been loaded successfully?
    Answer:
    It is most likely that this is a DataSource that does not send delta data to the BW System via the delta queue but directly via the extractor . You can display the current delta data for these DataSources using TA RSA3 (update mode ='D')
    Question 7:
    Do the entries in Table ROIDOCPRMS have an impact on the performance of the loading procedure from the delta queue?
    Answer:
    The impact is limited. If performance problems are related to the loading process from the delta queue, then refer to the application-specific notes (for example in the CO-PA area, in the logistics cockpit area, and so on).
    Caution: As of PlugIn 2000.2 patch 3, the entries in Table ROIDOCPRMS are as effective for the delta queue as for a full update. Note, however, that LUWs are not split during data loading for consistency reasons. This means that when very large LUWs are written to the delta queue, the actual package size may differ considerably from the MAXSIZE and MAXLINES parameters.
    Question 8:
    Why does it take so long to display the data in the delta queue (for example approximately 2 hours)?
    Answer:
    With PlugIn 2001.1 the display was changed: you are now able to define the amount of data to be displayed, to restrict it, to selectively choose the number of a data record, to make a distinction between the 'actual' delta data and the data intended for repetition, and so on.
    Question 9:
    What is the purpose of the function 'Delete Data and Meta Data in a Queue' in RSA7? What exactly is deleted?
    Answer:
    You should act with extreme caution when you use the delete function in the delta queue. It is comparable to deleting an InitDelta in the BW System and should preferably be executed there. Not only do you delete all data of this DataSource for the affected BW System, but you also lose all the information concerning the delta initialization. Then you can only request new deltas after another delta initialization.
    When you delete the data, this confirms the LUWs kept in the qRFC queue for the corresponding target system. Physical deletion only takes place in the qRFC outbound queue if there are no more references to the LUWs.
    The delete function is intended for example, for cases where the BW System, from which the delta initialization was originally executed, no longer exists or can no longer be accessed.
    Question 10:
    Why does it take so long to delete from the delta queue (for example half a day)?
    Answer:
    Import PlugIn 2000.2 patch 3. With this patch the performance during deletion improves considerably.
    Question 11:
    Why is the delta queue not updated when you start the V3 update in the logistics cockpit area?
    Answer:
    It is most likely that a delta initialization had not yet run or that the the delta initialization was not successful. A successful delta initialization (the corresponding request must have QM status 'green' in the BW System) is a prerequisite for the application data to be written to the delta queue.
    Question 12:
    What is the relationship between RSA7 and the qRFC monitor (Transaction SMQ1)?
    Answer:
    The qRFC monitor basically displays the same data as RSA7. The internal queue name must be used for selection on the initial screen of the qRFC monitor. This is made up of the prefix 'BW, the client and the short name of the DataSource. For DataSources whose name is shorter than 20 characters, the short name corresponds to the name of the DataSource. For DataSources whose name is longer than 19 characters (for delta-capable DataSources only possible as of PlugIn 2001.1) the short name is assigned in Table ROOSSHORTN.
    In the qRFC monitor you cannot distinguish between repeatable and new LUWs. Moreover, the data of a LUW is displayed in an unstructured manner there.
    Question 13:
    Why is there data in the delta queue although the V3 update has not yet been started?
    Answer:
    You posted data in the background. This means that the records are updated directly in the delta queue (RSA7). This happens in particular during automatic goods receipt posting (MRRS). There is no duplicate transfer of records to the BW system. See Note 417189.
    Question 14:
    Why does the 'Repeatable' button on the RSA7 data details screen not only show data loaded into BW during the last delta but also newly-added data, in other words, 'pure' delta records?
    Answer:
    It was programmed so that the request in repeat mode fetches both actually repeatable (old) data and new data from the source system.
    Question 15:
    I loaded several delta inits with various selections. For which one
    is the delta loaded?
    Answer:
    For delta, all selections made via delta inits are summed up. This
    means a delta for the 'total' of all delta initializations is loaded.
    Question 16:
    How many selections for delta inits are possible in the system?
    Answer:
    With simple selections (intervals without complicated join conditions or single values), you can make up to about 100 delta inits. It should not be more.
    With complicated selection conditions, it should be only up to 10-20 delta inits.
    Reason: With many selection conditions that are joined in a complicated way, too many 'where' lines are generated in the generated ABAP source code which may exceed the memory limit.
    Question 17:
    I intend to copy the source system, i.e. make a client copy. What will happen with may delta? Should I initialize again after that?
    Answer:
    Before you copy a source client or source system, make sure that your deltas have been fetched from the delta queue into BW and that no delta is pending. After the client copy, an inconsistency might occur between BW delta tables and the OLTP delta tables as described in Note 405943. After the client copy, Table ROOSPRMSC will probably be empty in the OLTP since this table is client-independent. After the system copy, the table will contain the entries with the old logical system name which are no longer useful for further delta loading from the new logical system. The delta must be initialized in any case since delta depends on both the BW system and the source system. Even if no dump 'MESSAGE_TYPE_X' occurs in BW when editing or creating an InfoPackage, you should expect that the delta has to be initialized after the copy.
    Question 18.
    Am I permitted to use the functions in Transaction SMQ1 to manually control processes?
    Answer:
    Use SMQ1 as an instrument for diagnosis and control only. Make changes to BW queues only after informing BW Support or only if this is explicitly requested in a note for Component 'BC-BW' or 'BW-WHM-SAPI'.
    Question 19.
    Despite the delta request only being started after completion of the collective run (V3 update), it does not contain all documents. Only another delta request loads the missing documents into BW. What is the cause for this "splitting"?
    Answer:
    The collective run submits the open V2 documents to the task handler for processing. The task handler processes them in one or several parallel update processes in an asynchronous way. For this reason, plan a sufficiently large "safety time window" between the end of the collective run in the source system and the start of the delta request in BW. An alternative solution where this problem does not occur is described in Note 505700.
    Question 20.
    Despite deleting the delta init, LUWs are still written into the DeltaQueue
    Answer:
    In general, delta initializations and deletions of delta inits should always be carried out at a time when no posting takes place. Otherwise, buffer problems may occur: If you started the internal mode at a time when the delta initialization was still active, you post data into the queue even though the initialization had been deleted in the meantime. This is the case in your system.
    Question 21.
    In SMQ1 (qRFC Monitor) I have status 'NOSEND'. In the Table TRFCQOUT, some entries have the status 'READY', others 'RECORDED'. ARFCSSTATE is 'READ'. What do these statuses mean? Which values in the field 'Status' mean what and which values are correct and which are alarming? Are the statuses BW-specific or generally valid in qRFC?
    Answer:
    Table TRFCQOUT and ARFCSSTATE: Status READ means that the record was read once either in a delta request or in a repetition of the delta request. However, this still does not mean that the record has successfully reached the BW. The status READY in the TRFCQOUT and RECORDED in the ARFCSSTATE means that the record has been written into the delta queue and will be loaded into the BW with the next delta request or a repetition of a delta. In any case only the statuses READ, READY and RECORDED in both tables are considered to be valid. The status EXECUTED in TRFCQOUT can occur temporarily. It is set before starting a delta extraction for all records with status READ present at that time. The records with status EXECUTED are usually deleted from the queue in packages within a delta request directly after setting the status before extracting a new delta. If you see such records, it means that either a process which confirms and deletes records loaded into the BW is successfully running at the moment, or, if the records remain in the table for a longer period of time with status EXECUTED, it is likely that there are problems with deleting the records which have already been successfully been loaded into the BW. In this state, no more deltas are loaded into the BW. Every other status indicates an error or an inconsistency. NOSEND in SMQ1 means nothing (see note 378903). However the value 'U' in field 'NOSEND' of table TRFCQOUT is of concern.
    Question 22.
    The extract structure was changed when the delta queue was empty. Afterwards new delta records were written to the delta queue. When loading the delta into the PSA, it shows that some fields were moved. The same result occurs when the contents of the delta queue are listed via the detail display. Why is the data displayed differently? What can be done?
    Answer:
    Make sure that the change of the extract structure is also reflected in the database and that all servers are synchronized. We recommend resetting the buffers using Transaction $SYNC. If the extract structure change is not communicated synchronously to the server where delta records are being created, the records are written with the old structure until the new structure has been generated. This may have disastrous consequences for the delta. When the problem occurs, the delta needs to be re-initialized.
    Question 23. How and where can I control whether a repeat delta is requested?
    Answer:
    Via the status of the last delta in the BW Request Monitor. If the request is RED, the next load will be of type 'Repeat'. If you need to repeat the last load for any reason, manually set the request in the monitor to red. For the contents of the repeat, see Question 14. Delta requests set to red when data is already updated lead to duplicate records in a subsequent repeat, if they have not already been deleted from the data targets concerned.
    Question 24.
    As of PI 2003.1, the Logistic Cockpit offers various types of update methods. Which update method is recommended in logistics? According to which criteria should the decision be made? How can I choose an update method in logistics?
    Answer:
    See the recommendation in Note 505700.
    Question 25.
    Are there particular recommendations regarding the maximum data volume of the delta queue to avoid danger of a read failure due to memory problems?
    Answer:
    There is no strict limit (except for the restricted number area of the 24-digit QCOUNT counter in the LUW management table - which is of no practical importance, however - or the restrictions regarding the volume and number of records in a database table).
    When estimating "soft" limits, both the number of LUWs and the average data volume per LUW are important. As a rule, we recommend bundling data (usually documents) as soon as you write to the delta queue to keep number of LUWs low (this can partly be set in the applications, for example in the Logistics Cockpit). The data volume of a single LUW should not be much larger than 10% of the memory available to the work process for data extraction (in a 32-bit architecture with a memory volume of about 1 GByte per work process, 100 MByte per LUW should not be exceeded). This limit is of rather small practical importance as well since a comparable limit already applies when writing to the delta queue. If the limit is observed, correct reading is guaranteed in most cases.
    If the number of LUWs cannot be reduced by bundling application transactions, you should at least make sure that the data is fetched from all connected BWs as quickly as possible. But for other, BW-specific, reasons, the frequency should not exceed one delta request per hour.
    To avoid memory problems, a program-internal limit ensures that no more than 1 million LUWs are ever read and fetched from the database per delta request. If this limit is reached within a request, the delta queue must be emptied by several successive delta requests. We recommend, however, to try not to reach that limit but trigger the fetching of data from the connected BWs as soon as the number of LUWs reaches a 5-digit value.
      THANKS =POINTS in SDN
    SANJEEV

  • Deleting data from a compressed cube

    Hi Gurus,
    I have an info cube which is getting data fron three different DSO's (DSO1, DSO2, DSO3)
    and I have Aggregates to the cube and this whole process is automated in a Process chain
    now I have to delete the data in the cube which is getting loaded from DSO2 only
    how to achieve this
    Thanks
    MSS:-)

    Hi,
    You can do Selective deletion but how you do that , is there any field which differentiate between DSO1 , DSO2 & DSO3
    based on that fields you make selective deletion
    for Dynamic Selective deletion goto the following link
    http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/4001f4b1-189f-2e10-bcb3-87bb0fb428ba
    another option is to create infoset/ multiprovider  based on DSO1 & DSO3
    Best Regards
    Obaid

  • Error while creating Info Spoke....

    Hi,
    I am facing one problem while creating Info Spoke.I have 4 DSO which included in one Infoset & that infoset incluede in one Multiprovider. I want to creat info spoke on this Multiprovider.But when i changed the tab or at time of save & activate it gaves me error as "InfoObject ZCGNCRM2___F100 is not available in version A". I checked this info-obj.its already in active version. I reactivate it.I reactivate infoset & multiprovider also.But still exist this error.
    Please suggest me any solutuion.

    Hi Sandeep,
    Check this info object is included in any of your DSO or Info set and Multi provider.If yes check all are active or not.
    activate the info object once again using program RSDG_IOBJ_ACTIVATE and activate your Info spoke
    Check below article
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/f0d48577-9542-2e10-05a1-d6c1e07ea641?quicklink=index&overridelayout=true
    Regards,
    Venkatesh

  • SAP BI Research Study?

    Hello Everyone, I am currently pursuing my MBA (MIS) at California State University, Chico. Since a couple of weeks I have been searching for some topic in the domain of SAP BI for conducting an independent research study. I am unable to figure out any solution for the same.
    Over the last semester I have gained tremendous insights on SAP Business Intelligence via. a grad level course. I did get hands on experience on data warehousing as well. But after I began my search on the Internet to familiarize myself with the current situation I was flabbergasted. There is so much happening around us and there are a plethora of skills to acquire that I feel like robbing a bank to become a millionaire only to enroll myself for all the courses across the globe which would inundate me with cutting edge skills within SAP. I consider myself as a toddler. Actually, not even a toddler! But I do not intend to get demotivated and neither do I want to be a part of the problem. I need to find solutions.
    So I would like all you experts to pitch in your ideas as to what you feel are considerably the most sustainable topics in SAP BI and Data Warehousing? Which topics need more research hours and would demand oneself to ponder upon it with an intent to come up with innovations later in our research? Which topics have the most potential? (By topics I mean: Either related to warehousing...DTP's, Transformations, Delta, Process Chains, Cubes, DSO's, Queries ..MDM in general..etc etc etc. or say SAP BI Accelerator, SAP BO Explorer, SOA, BPM, Portal etc)
    With regards to the aforementioned topic set I would also want to host a question: After you suggest a topic out of the above mentioned set, would you all still think it is going to be worthwhile spending my time on such topics or would you suggest me to look out for something new in the BI world which may help me at work later or even embellish my resume...for e.g. working on some innovation or some fancy idea (with out without a programmer alongside me) I may sound obtuse but would the following type of study help: "Considering the need for better analysis of transactional data by anyone in general, not necessarily the top-level so to speak, what if we have audio files (results) (podcasts sorts) for iphones if the display is not large enough to fit graphs, bar charts etc. One can always have a wireless bluetooth dead set device for constantly having an option to avail days end results for critical issues." This is what I feel is fancy. Could be immature too. But would this help? I may have sounded immature but I need to figure out the right path..which is why I have come to the right bunch of people.
    I am awaiting some of the most informative responses. Please acknowledge!
    Thanks and Regards, Akshay Thussu

    Hello,
    Most sustainable topics in SAP BI and Data Warehousing?
    Yes, Data warehousing is the most important area where u can learn basic concept that is required for SAP BI.
    Which topics need more research hours and would demand oneself to ponder upon it with an intent to come up with innovations later in our research?
    Scenario wise, Data Modelling, ETL (i.e Extraction, Transformation & Loading) & then Reporting are the topics that u can go 1 by 1.
    Which topics have the most potential? (By topics I mean: Either related to warehousing...DTP's, Transformations, Delta, Process Chains, Cubes, DSO's, Queries ..MDM in general..etc etc etc. or say SAP BI Accelerator, SAP BO Explorer, SOA, BPM, Portal etc)
    Now, as u mentioned in above question Cube, DSO's etc. are all the contents which is mainly used for data modelling & Reporting.
    Ok, so following are the links that will help u so thru the links,
    Data warehousing workbench:
    http://help.sap.com/saphelp_nw2004s/helpdata/en/a8/6b023b6069d22ee10000000a11402f/content.htm
    Infoproviders
    http://help.sap.com/saphelp_nw04/helpdata/EN/02/bafc3f9ec4e669e10000000a155106/content.htm
    Infoobjects
    http://help.sap.com/saphelp_nw70/helpdata/EN/80/1a63f9e07211d2acb80000e829fbfe/frameset.htm
    Infoproviders
    Different types of Infoproviders are : Infocube,DSO,Virtual providers,Infoset,Multiprovider...in this link you will find all the details..
    http://help.sap.com/saphelp_nw70/helpdata/EN/80/1a63f9e07211d2acb80000e829fbfe/frameset.htm
    Master Data
    http://help.sap.com/saphelp_nw70/helpdata/EN/04/cce63736544d4de10000009b38f8cf/frameset.htm
    Modelling,Transformation,Furthur Processing Data,Data Distribution,Real Time Data Aquisition.....all these topics you will find in the following link:
    http://help.sap.com/saphelp_nw70/helpdata/EN/f8/7913426e48db2ce10000000a1550b0/frameset.htm
    Transferring Data from flat files
    http://help.sap.com/saphelp_nw70/helpdata/EN/43/03450525ee517be10000000a1553f6/frameset.htm
    Data Aquisition
    http://help.sap.com/saphelp_nw70/helpdata/EN/80/1a618ae07211d2acb80000e829fbfe/frameset.htm
    Infopackages
    http://help.sap.com/saphelp_nw70/helpdata/EN/80/1a6574e07211d2acb80000e829fbfe/frameset.htm
    *Data Transfer Process *
    http://help.sap.com/saphelp_nw70/helpdata/EN/42/fa50e40f501a77e10000000a422035/frameset.htm
    Aggregates
    http://help.sap.com/saphelp_nw70/helpdata/EN/26/4bc0417951d117e10000000a155106/frameset.htm
    Restricted and Calculated Keyfigure
    http://help.sap.com/saphelp_nw70/helpdata/EN/b4/730276938a504cacc56b44fd3578e3/frameset.htm
    For Bex Query,Reporting,Bex Analyzer
    http://help.sap.com/saphelp_nw70/helpdata/EN/80/3424429e905033e10000000a155106/frameset.htm
    BEx Web Analyzer :
    http://help.sap.com/saphelp_nw70/helpdata/en/0d/af12403dbedd5fe10000000a155106/frameset.htm
    BI Content
    http://help.sap.com/saphelp_nw70/helpdata/EN/37/5fb13cd0500255e10000000a114084/frameset.htm
    Most Imp Link
    http://help.sap.com/saphelp_nw70/helpdata/EN/46/8dfebaebf268e5e10000000a1553f7/frameset.htm
    Hope it will helps u for ur research,
    Best Regards,
    Santosh Dhanawade

  • Details about Data Target

    Details about Data Target

    Hi...........
    Infoproviders
    http://help.sap.com/saphelp_nw04/helpdata/EN/02/bafc3f9ec4e669e10000000a155106/content.htm
    Infoobjects
    http://help.sap.com/saphelp_nw70/helpdata/EN/80/1a63f9e07211d2acb80000e829fbfe/frameset.htm
    Infoproviders
    Different types of Infoproviders are : Infocube,DSO,Virtual providers,Infoset,Multiprovider...in this link you will find all the details..
    http://help.sap.com/saphelp_nw70/helpdata/EN/80/1a63f9e07211d2acb80000e829fbfe/frameset.htm
    Master Data
    http://help.sap.com/saphelp_nw70/helpdata/EN/04/cce63736544d4de10000009b38f8cf/frameset.htm
    Hope this helps you............
    Regards,
    Debjani.............

  • Impact on Master data Infoobject

    Hi Bwers,
    Currently we are making changes to exsiting master data infoobject (Business partner) in the system through adding one more Attribute. And we are including 2 new info objects to exsiting DSO also.
    I was asked to investigate impact of these chanes in systems. Pls provide your valuable suggestions about this activity.
    1.  What about the impact on other info objects (say 0CUSTMOER) which is reference to Bussiness partner.
    2. If any infoset incluing this infoobject, what is the impact on this Infoset / Multiprovider.
    3. If Business partner has been used as a navigational attribute for any other Infoobjects.
    Impact on DSO:-
    1. Impact on above data targets.
    2. And transformation between DSO to above all the infoproviders.
    How to investigate this things in the system.
    Regards,
    Venkat

    1. What about the impact on other info objects (say 0CUSTMOER) which is reference to Bussiness partner.
    > Adding an Attribute to wont affwct 0Customer.
    2. If any infoset incluing this infoobject, what is the impact on this Infoset / Multiprovider.
    > You need to adjust your inoset/MP. In case you want this attribute in the report, you have to select this object in infoset/MP.
    3. If Business partner has been used as a navigational attribute for any other Infoobjects.
    > It can still be used like that.
    Impact on DSO:-
    1. Impact on above data targets.
    ---> need to activate the DSO and load the data again. Need to change the transformation as well.
    2. And transformation between DSO to above all the infoproviders.
    > Those will get inactive. You have to activate it again.
    Thanks...
    Shambhu

  • Universe Turns Bex Date fields to Character

    We are noticing that the universes that we creat are turning all date fields from the bex query to charachter fields. When we try to change the field back to a date in the universe it fails saying that the "driver is not capable". This is usually tied to bad data in the columns or a date format issue. Has anyone seen this and come up with a workaround?

    Hi Ingo,
    We r facing the same problem with date fields.
    The universe is built on SAP BI Query, the data model is as follows:
    DSOs> Infoset-> Multiprovider->SAP BI Query->Universe
    There are  Date variables (DATS Type) defined in the SAP BI query.which are not appearing in calander format in WEBI report at selection prompts screen, rather it is displaying as list of values.
    I checked at the universe level the date dimension for that variable is reflecting as character data type instead of date data type.
    I dont find any key detail object under date dimension whose data type is DATE.
    Edited by: Nisha Makhija on Jul 17, 2009 8:39 PM

  • Some question about the query designer

    hello, dear all,
    I am a new comer of here, and I am intersting in BI, but I have no basic knowledge about it.
    so I just want someone could give me some advice about it.
    our boss need I do the developer about the query designer,  I just have searched in this forum. but nothing founded for I am a new comer here,
    I heard there are some training document of the query designer, could someone give me the URL, thanks.

    Hi,
    Query desinger is used to develop a Query, Query can be created on following data targets
    -Info Cube
    -DSO
    Virtual data target
    -MultiCube
    -Infoset
    -Multiprovider
    We have 5 section in query designer
    - Infoprovider : where we select the data target , on which report to be created
    -Filter : to restrict value at infoprovider level ( if you want data for year 2008, for example)
    -Free Characterstic : this allow you to drill down
    -Columns : char/keyfigs to be display in columns can be added here
    Row: key/char to be display in Rows can be added here
    gv me ur mailid i will let u have Bex manual,
    I would suggest , if you have any IDES ( training ) system , where you can log in and then go to RRMX,
    and try to create new query and add any data target ( which is already created ) and then drag and drop the char/key fig to the required section ,
    save it and execute it .....
    if you play arround and see the output , that would help u to understand how to work with query designer.
    Hope this helps
    Sukhi
    Edited by: Sukhvidner Singh on Nov 4, 2009 5:36 PM

  • Difference between InfoSet and Multiprovider

    Hello Gurus,
    Could someone tell me the difference between an InfoSet and a Multiprovider? And in what occasions we should use these?
    Thank you.
    DM

    Hi David,
    1. Infoset and Multi provider physically doesn't have any data.
    2. Where they are used?
    Ans: they are used for reporting.
    3. Why they are used?
    Ans: when we want to see a report on a data coming from two or more different tables then we go for infoset or multiprovider.
    Example: suppose if there are 2 tables A which contains ( customer id, customer name..etc )
                   and table B contains ( customer id, material id, material name..etc )
    -> if we want a report on cust id , cust name , mat id , mat name in a single report then we go for MULTIPROVIDER ( since multi provider can fetches the data from two fileds )
    -> if we want a report on cust id from both tables and cust name from table A then we can go for INFOSET ( since info set fetches the data which is common in both tables and the other data from any table as per req. by using left outer join condition )
    So we can go for any in this as per our requirement.
    One more thing is that INFOSET can be done on Info Cube and DSO only.
    MULTIPROVIDER can be done on data source, PSA , DSO, data target..etc
    hope this helps u, and let me know if u need any further info.
    Thanks in advance
    Sai Chand.

Maybe you are looking for