Modeling the Data model when multiple source systems

Hi gurus,
i have a scenario where i am getting data from multiple source systems all R/3 but from different locatins.
my doubt is if i have to do reporting on the data available.
1. do i have to build seperate infocubes for each source systems or build seperate data designs for each source system .
2. How can i consolidate the different source sytems data into one and report sensibly as well as without loosing the source system identification.
thanks and regards
Neel

Hi all,
thanks for your focus, ya i am also thinking of have a flexible solution where in i can do it easily as well as we don't have much of complexcity, but wanted to try different options as well so that we can go for best one.
I thought of multiprovider at first when RafB suggested, and i agree with you as well Lilly, that data volume will be a problem, so keeping all this in view i want to build a solution where in it will be sensible as well as not confusing in reports ( i mean clear and readable reports)
[email protected]
please kindly forward any documents which might be helpful for me in this scenario
thanks and regards
neel
Message was edited by: Neel Kamal

Similar Messages

  • When we are able to see the data in Tcode RSA3(Source system)

    Hi,
    Please help me!
    In enhancement after doing the R/3 stuff, when we are able to see the data in Tcode RSA3 in R/3 only?
    Give the detail for all types of data source and for all types of loads.
    eg: Std data source for deltaload, Generic data source for full load.....etc...
    Regards,
    Shiva.

    Sure...
    I am doing enhancement for datasource '2LIS_05_QVUDN'.
    I complited the R/3 stuff and check the data for enhanced fields in RSA3...it's giving run timeout run time error.
    Then I just ignore that error and I complited BW stuff successfully....after compliting BW stuff I am able to see the data in RSA3 in R/3.
    So my question is, What is the root cause for this....?
    When we are able to see the data in RSA3 in source system.?
    please give me details for all types of datasource and all types of loads so that it will helps a lot for me in future.

  • SLT Replication for the same table from Multiple Source Systems

    Hello,
    With HANA 1.0 SP03, it is now possible to connect multiple source systems to the same SLT Replication server and from there on to the same schema in SAP HANA - does this mean same table as well? Or will it be different tables?
    My doubt:
    Consider i am replicating the information from KNA1 from 2 Source Systems - say SourceA and SourceB.
    If I have different records in SourceA.KNA1 and SourceB.KNA1, i believe the records will be added during the replication and as a result, the final table has 2 different records.
    Now, if the same record appears in the KNA1 tables from both the sources, the final table should hold only 1 record.
    Also, if the same Customer's record is posted in both the systems with different values, it should add the records.
    How does HANA have a check in this situation?
    Please throw some light on this.

    Hi Vishal,
    i suggest you to take a look to SAP HANA SPS03 Master Guide. There is a comparison table for the three replication technologies available (see page 25).
    For Multi-System Support, there are these values:
    - Trigger-Based Replication (SLT Replication): Multiple source systems to multiple SAP  HANA instances (one  source system can be connected to one SAP HANA schema only)
    So i think that in your case you should consider BO Data Services (losing real-time analytics capabilities of course).
    Regards
    Leopoldo Capasso

  • Error in the Data Selection - Errors in Source System RSM 340

    Hey Everyone,
    I'm having an issue with loading a Delta for 0MAT_PLANT_ATTR.  This is the second night it has happened, where I get the error "Error in the Data Selection".  This error happens in the EXTRACTION part of the Monitor.  And, we cannot run another Delta to see if it'll work, we have to delete the old Delta Init, and run a new one.  The other error is "Errors in Source System - RSM340".  That's as much it gives.
    Could anyone suggest why this would be happening?  I checked RSA3 in R/3, and no errors on the extraction.  I also checked sm37 against the job that loads that data, and it completed successfully.  I honestly don't think there's an issue with the data.  We cannot have this happen often, and I just need to know if this is something that can be corrected.
    Thanks,
    John

    There is nothing wrong with the TID's or anything like that.  There is also no selection criteria in the infopackage.
    When the Delta fails, the error occurs under EXTRACTION in the monitor.  It seems to be successful under TRANSFER and PROCESSING in the monitor, which I find strange.  Also, I run a new Delta Init with Data Transfer, and that works just fine.
    So, I'm not sure what it could be?

  • How to Load Master Data Text  from Multiple Source Systems

    Situation:  Loading vendor master (text) from two systems.  Some of the vendor keys/number are the same and so is the description.  I understand that if I was loading Attributes, I could create another attribute such a source system id and compound it to 0Vendor.  However, I don't believe you can do the same when loading text. 
    Question:  So, how can I load duplicate vendor keys and description into BW using the *_TEXT datasource so they do not overwrite?
    Thanks!

    Hi,
    Don't doubt, it'll work!
    Sys ID as compound to 0Vendor.
    If you doubt about predefined structure of the *_TEXT datasource and direct infosource, then use a flexible infosource. Maybe you'll have to derive the sys ID in a routine or a formula.
    Best regards,
    Eugene

  • Use a multiple selection parameter in the data model

    Hi
    i have created a parameter with the property multiple selection.
    This parameter i use in the Data Model where area,
    if i just mark one value the report works well if i mark more the one
    the report run into a error.
    WHERE PSNR IN :ma
    the xdo looks like
    <?xml version = '1.0' encoding = 'utf-8'?>
    <report version="1.1" xmlns="http://xmlns.oracle.com/oxp/xmlp" defaultDataSourceRef="tisoware">
    <title>BDE Nettozeit</title>
    <properties>
    <property name="showControls" value="true"/>
    <property name="online" value="true"/>
    <property name="parameterColumns" value="2"/>
    <property name="openLinkInNewWindow" value="true"/>
    </properties>
    <dataModel defaultDataSet="BDERUECK">
    <dataSet id="BDERUECK">
    <sql dataSourceRef="tisoware">
    <![CDATA[SELECT (SUBSTR(DATUM,1,6)) AS DATUM, SUM(NTTOZEIT) AS ZEIT,
    TRUNC(SUM(NTTOZEIT)/60) AS STUNDE
    FROM BDERUECK
    WHERE (DATUM  >= :datev) AND (DATUM <= :dateb) AND (KZDELETE = 'N') AND FIRMA = '1' AND ((PSNR IN :ma))
    GROUP BY (SUBSTR(DATUM,1,6))]]>
    </sql>
    <input id="datev" value="${datev}" dataType="xsd:string"/>
    <input id="dateb" value="${dateb}" dataType="xsd:string"/>
    <input id="ma" value="${ma}" dataType="xsd:string"/>
    </dataSet>
    </dataModel>
    <valueSets>
    <valueSet id="date">
    <sql dataSourceRef="tisoware">
    <![CDATA[SELECT DISTINCT DATUM
    FROM BDERUECK]]>
    </sql>
    </valueSet>
    <valueSet id="employee">
    <sql dataSourceRef="tisoware">
    <![CDATA[select      PERSTAMM.PSPERSNR as PSPERSNR,
          PERSTAMM.PSNR as PSNR
    from      TISOWARE.PERSTAMM PERSTAMM
    where    PERSTAMM.FIRMA = '1'
    order by PERSTAMM.PSPERSNR]]>
    </sql>
    </valueSet>
    </valueSets>
    <parameters>
    <parameter id="datev" defaultValue="20050103">
    <select label="Datum von : " valueSet="date" multiple="false" all="false" refreshParameters="false"/>
    </parameter>
    <parameter id="dateb" defaultValue="20060930">
    <select label="Datum bis : " valueSet="date" multiple="false" all="false"/>
    </parameter>
    <parameter id="ma" defaultValue="152">
    <select label="Mitarbeiter : " valueSet="employee" multiple="true" all="false"/>
    </parameter>
    </parameters>
    <templates default="BDE Nettozeit">
    <template label="BDE Nettozeit" type="rtf" url="BDE Nettozeit.rtf"/>
    </templates>
    </report>

    Alex,
    the missing thing in your example is the fact, that if only one value is selected, the parameter has exact this value like BOSTON. If you choose more than one value, the parameter includes the *'*, so that it looks like *'BOSTON','NEW YORK'*. So you need to check in the package, if there's a *,* in the parameter or not. If yes there's more than one value, if not it's only one value or it's null.
    So change your package to (you need to expand your variables)
    create or replace package bip_departments_2_parameters
    as
    p_dep_2_param varchar2(1000);
    p_loc_1_param varchar2(1000);
    p_where_clause varchar2(1000);
    function beforereporttrigger
    return boolean;
    end bip_departments_2_parameters;
    create or replace package body bip_departments_2_parameters
    as
    function beforereporttrigger
    return boolean
    is
    l_return boolean := true;
    begin
    p_where_clause := ' ';
    if p_dep_2_param is not null then
    if instr(p_dep_2_param,',')>0 then
    p_where_clause := 'WHERE DNAME in ('||p_dep_2_param||')';
    else
    p_where_clause := 'WHERE DNAME = '''||p_dep_2_param||'''';
    end if;
    if p_loc_1_param is not null then
    if instr(p_loc_1_param,',')>0 then
    p_where_clause := p_where_clause || ' AND LOC IN ('||p_loc_1_param||')';
    else
    p_where_clause := p_where_clause || ' AND LOC = '''||p_loc_1_param||'''';
    end if;
    end if;
    else
    if p_loc_1_param is not null then
    if instr(p_loc_1_param,',')>0 then
    p_where_clause := p_where_clause || 'WHERE LOC in ('||p_loc_1_param||')';
    else
    p_where_clause := p_where_clause || 'WHERE LOC = '''||p_loc_1_param||'''';
    end if;
    end if;
    end if;
    return (l_return);
    end beforereporttrigger;
    end bip_departments_2_parameters;
    I've written a similar example at http://www.oracle.com/global/de/community/bip/tipps/Dynamische_Queries/index.html ... but it's in german.
    Regards
    Rainer

  • Excel 2013 PowerPivot Error - "PowerPivot is unable to load the Data Model"

    I'm attempting to use the PowerPivot add-on in Excel.  When clicking 'Manage Data Model' or attempting to 'Load to data model,' I receive the error 'We couldn't load the Data Model.  This may be because the Data Model in this workbook is damaged.',
    followed by 'PowerPivot is unable to load the Data Model.'  This happens irrespective of data source type.  I have Excel 2013 32-bit, PowerPivot 32-bit, running on Windows 7 OS 64-bit.  I am not running an SSAS instance on my machine. 
    Any suggestions?

    Hi
    Have you ever tried to repair your office?
    Also we may refer to the following blog :
    http://blogs.technet.com/b/the_microsoft_excel_support_team_blog/archive/2013/11/12/powerpivot-for-excel-2013-errors-after-october-update-kb-2825655.aspx
    This issue seems to be caused by the October 2013 update (KB 2825655) for Excel 2013. Try to remove it or restore the system to an earlier point and check the result.
    Regards
    Tylor Wang
    TechNet Community Support

  • PowerPivot removed table replicated into the data model on refresh

    Hey
    I am facing the following problem which is absolutely persistent and annoying in nature. I am using a data model created with PowerPivot with connections to three different SQL servers and 25 tables. The bug is that every time I refresh my data, a table
    that was initially added to the data model (and later removed) is replicated in equivalent number of tables that use this connection, which is currently around 12 tables, so I get 12 duplicates of the same unnecessary table on every refresh, named with the
    name of the schema on the SQL server that is followed by a running number. Removing these from the data model on every refresh simply takes too much time on a daily basis. The additional undesired table which is created is what is defined in the SQL query
    of the connection string of this connection.
    I tried to solve this problem by zipping the Excel file and modifying this SQL statement present in the connection string in the connections.xml file in a way that it would not return anything, but this always leads into a corrupted workbook.
    After encountering this I have managed to avoid this problem by making sure that the initial table I'm adding to the data model will remain in the model, but I really don't want to rebuild this current model. Is there any fix for this issue without
    rebuilding the data model?
    Thanks for all the help in advance!

    I am using 64-bit Office 2013 with the recent SP1 upgrade and the included PowerPivot plug-in on a Windows 8.1 machine.
    I always create the connection by using "From SQL server" then fill in the server and log in details and use a ready SQL statement that I have written by using SQL Server Management Studio to import the initial table. For all the preceding tables
    using the same server and schema, I always select it from existing connections and just paste the SQL queries required for any additional tables to the "Write a query that will specify the data to import", rather than using the table import tool.
    I am not sure if it could have any effect that I've been using the type connection "From SQL server" although the database is really running on Azure (there is a possibility to select From Other sources > Microsoft SQL Azure, but the functionality
    seems the same).
    I don't really understand what I should edit from the existing connections section. Just to clarify, apparently the Connection String parameters can be edited from PowerPivot, but not the Command Text which can be viewed from Excel Data > Connections
    > Properties, but is shown in gray with a message "Some properties cannot be changed because this connection was modified using the PowerPivot add-in."
    In the data model I am using two sets of identical tables, but connected to two different servers that are running a database 1:1 in structure. Initially the problem considered only the other of these, but I accidentally reproduced it by, again, editing
    the SQL query in the table that was initially imported when that connection was created (PowerPivot > Design > Table Properties). Now when refreshing I get tables that are structured as in what the "Connection Text" part of the connection would
    produce, which replicated for the equivalent amount of additional tables using the same connection, so now I'm getting 12 tables (the number of tables using the same connection) each named "*schemaname1* Query", "*schemaname1* Query2",
    "*schemaname1* Query3" "*schemaname2* Query" and so on.
    Personally I can definitely see a pattern here. If there isn't a table matching what has been specified in the "Command Text" that was described when the connection was created, then it for some reason runs this query anyway on every table that
    is using the connection.

  • Powerpivot Error on Refresh -- "We couldn't get data from the data model..."

    I'm using Excel 2013 and Windows 8.1.  I have a spreadsheet I've been using for over a year, and I've just started getting this error message when I try to refresh the data.
    "We couldn't get data from the Data Model.  Here's the error message we got:
    The 'attributeRelationship' with 'AttributeID' - 'PuttDistCat9' doesn't exist in the collection"
    Any idea how I can fix this problem?  I haven't changed anything related to that particular attribute.  All the data is contained in separate sheets in the workbook, so there are no external sources of data.
    Thanks.
    Jean

    Thanks for all the suggestions.
    I found a slightly older version of the spreadsheet that still refreshes properly, so I don't think I have any issues with the version of Excel or Power Query.  (I've had this same error before, and I believe I applied the hotfix at that time.)
    I think this problem started after I updated a number of the date filters in the pivot tables.  I haven't made any changes to the data model, and the only updates I've made were to add data (which I do all the time), and to change the date filters on
    the pivot tables.
    As suggested, I added a new pivot table querying one table (the table with the attribute that shows up in the error message), and it worked fine.  I can also refresh this pivot table.
    Then I tried adding a pivot table which went against several tables in the data model (including the table in question).  The pivot table seemed to return that data properly.  However, when I tried to refresh it, I got the same error message ("we
    couldn't get data from the data model..."). 
    Dany also suggested running the queries one at a time to see which one is in error.  Without checking all the pivot tables, it appears that any which use the table "HolePlayedStrokes" generate the error (this is the table with the attribute
    mentioned in the error message).  Pivot Tables without that particular table seem to refresh OK.  Unfortunately, that is the main table in my data model, so most of the pivot tables use it.
    Any other suggestions?  I'd be happy to send a copy of the spreadsheet.
    Thanks for all the help.
    Jean

  • BI Publisher 10g - Limit to the size of the data model?

    Hello,
    We are using OBIEE 10g 10.1.3.4.1 in our organization and have a problem when working on a large report in BI Publisher. It seems that once our data model gets large enough, we are no longer able to edit the report. When we click on the Edit link to load the report editor, the side menu that lists the data model, parameters, bursting, etc does not load, preventing us from making any changes to the report.
    At first we thought it was a timeout issue and changed the following:
    $J2EE_HOME/applications/xmlpserver/xmlpserver/js/XDOReportModel.js
    Property name : XDOReportModel.TIMEOUTMSEC
    set value in Milli seconds as " XDOReportModel.TIMEOUTMSEC = 15000
    After making the change and restarting the services, it did not help. Is there a different place for the timeout of loading the report editor? What about some kind of cap to the size of the data model? We have one large data template that is used to generate a report. I tried breaking the large data set into multiple, smaller data sets (still using data tempalte) and that didn't seem to help any. Is there a limit the size of the data model BI Publisher can return?
    Any help would be greatly appreciated! Thanks!

    Hello,
    I think you should create a recordStore with RMS and use it like a string array. You can store huge amount of data and the time access when request data from recordStore is very similar as if it is store on the RAM memory with recent devices.
    And if it is not persistent data you can delete the recordStore when closing the midlet.

  • Powerview "Add this to the Data Model" checkbox is greyed out.

    How can this checkbox "Add this to the Data Model" in the Import Data dialog box be made available and not greyed out?

    Hi Rufino,
    Data Model is a new approach for integrating data from multiple tables, effectively building a relational data source inside the Excel workbook. Within Excel, Data Models are used transparently, providing data used in PivotTables, PivotCharts, and Power
    View reports. between the data sources.
    In your scenario, it seems that you import data from Analysis Services database. However, this option is not available for Analysis Services database. So in you want to made this option, you can import data from relational database.
    Reference:Introduction to the Data Model and Relationships in Excel 2013
    Regards,
    Charlie Liao
    If you have any feedback on our support, please click
    here
    Charlie Liao
    TechNet Community Support

  • Importing editable data that can be compared to the data model?

    I will describe my objective and how I've solved it currently without Powerpivot.  Then I will ask if there's a way to improve upon my solution by using Powerpivot.
    The topic is project allocation forecasting.  Currently I import two copies of the same dataset from a work management system in to Excel.  One copy is then modified iteratively and produces graphs used in predict resources on various projects
    in the future.  We work through many "what if" modifications to this data.  Once we believe we have the data correct, I then compare the modified data to the second copy of the data, to see what we changed.
    With Powerpivot, much of what I coded in VBA for Excel previously is extremely simple (and better.)  Additional advantage is I don't have to actually import the data in to Excel (after first exporting it from the work management system.)  But I
    lost the ability to modify the data over and over, and then compare it back to its original source.
    Finally to my question... is there a better way to import the data in a way that it will be editable, and then compare that data back to the data existing in the data model?

    well, you cannot modify any data once it is loaded to power pivot - unfortunately
    the only thing i could think of is to use Excel linked tables which are then updated in power pivot once you modified them
    depending on how you want to change your data this may be an option for you
    Gerhard Brueckl
    blogging @ http://blog.gbrueckl.at
    working @ http://www.pmOne.com

  • "An error occured while working on the Data Model in the workbook" on some workbooks published to Power BI site

    Hello,
    I am using the Power BI for Office 365, and I have published several Excel 2013 workbooks having Power Pivot Data Models.
    I have a problem on some of the workbooks, once a slicer is selected, I get the error: "An error occurred while working on the Data Model in the workbook" and the slicers do not affect the charts.
    Some workbooks work perfectly fine. I am using the same user for all workbooks when creating and publishing. I tried with small workbooks less than < 10 MB size and larger workbooks > 10 MB. There is no rule, some workbooks larger than 10 MB work perfectly
    with the slicers effecting the charts, and some don't. Similarly for smaller size.
    Any ideas of how I can debug the cause of the issue? 
    Appreciate any feedback, 
    Thanks,
    Grace

    Hi Grace,
    I assume that the experience in the Excel client is working fine, right?
    Are you getting a correlation id with the error?
    Please send us a bit more information / samples to reproduce over email to
    this address.
    thanks,
    Guy
    GALROY

  • Why do I get the symbol data link in the query_name of the data model

    If I use an inline subquery within the SELECT of a data model query, the symbol data link appears in the query_name of the data model.
    This causes the data link between two queries not to be used even if there is an existing link between the two. The trace file doesn't
    have that join condition and so the child query in which the inline subquery was added fetches many redundant data.
    Any idea what that symbol represents and how to get rid of it. It also appears during other times, not only when we use inline subquery
    within a SELECT. Not sure if it can cause any harm to the report.
    The problem is seen in Report Builder 6.0.8.19.0. Not sure if the problem is with other releases also.
    Any suggestions are welcome.

    I am living with the symbol in the query box as well, because my report only has a single query. My query also has several inline subqueries. I had never seen this before last week, spent a day re-reading the manuals and couldn't figure it out.
    My post is not a solution, but a "me too" echo that could use an answer.
    Troy Tremble
    If I use an inline subquery within the SELECT of a data model query, the symbol data link appears in the query_name of the data model.
    This causes the data link between two queries not to be used even if there is an existing link between the two. The trace file doesn't
    have that join condition and so the child query in which the inline subquery was added fetches many redundant data.
    Any idea what that symbol represents and how to get rid of it. It also appears during other times, not only when we use inline subquery
    within a SELECT. Not sure if it can cause any harm to the report.
    The problem is seen in Report Builder 6.0.8.19.0. Not sure if the problem is with other releases also.
    Any suggestions are welcome.

  • Is it proper for an application module has 300 views in the data model?

    Based on my understanding, all view object instances are initialized at the same time as the application module.
    It follows that when an application module is associated with a user session, the queries of each view object in the data model are executed.
    ie. if the data model has 325 view object instances, 325 queries will be executed.
    This doesn't seem efficient:
    1. As each user session only need to access a subset of the view objects (for examples 35 views), the rest of view objects need not be instantiated for that session.
    2. Performance may be slow on executing all 325 queries at one time, even worse when some queries are complex queries.
    Am I correct on this?
    In this case, will it be better to instantiate the view objects using the view object definition at runtime instead of adding all of them to the application module at design time?

    No, that's not the default behaviour.
    Assuming you're talking JDev 11, the AM on it's General tab as a Tuning option where you can set the loading of the VOs to Lazy or Non-Lazy, Lazy is the default. In turn each VO on their General tab also has a Tuning option of similar nature.
    Online documentation says:
    Lazy LoadingSelect to choose lazy loading. In this case, the application module instance is created without any of its components. Your view object and nested application module instances will not load until the first time they are used. For example, when lazy loading is enabled, a view object is instantiated only when the application accesses it. This will make start-up time faster.
    Non-lazy Loading
    Select to choose non-lazy loading. Your view object and nested application module instances will load when the application starts up. For example, suppose the application module you are editing defines three view objects and two nested application modules (which appear in the Application Module Instance section of the Data Model page over the application module overview editor). When lazy loading is disabled, the application module instance is created together with all three view objects and both nested application modules.>
    So in essence as you describe for each session, if we're using the Lazy setting, even though you have 325 VOs within an AM, if the user within the session only visits 35 queries, only 35 queries should fire.
    Does this help?
    CM.

Maybe you are looking for

  • Text Caption Dialog window is black-help!

    For some reason, the text caption dialog box window is black. It did not used to be black, but somehow decided it wanted to be black now. This makes is incredibly difficult to view the black text as I attempt to edit it. Anyone else have this problem

  • InfoSource copied from 0BPARTNER_ATTR is not found in ABAP dictionary

    Hello friends, I am migrating 0BPARTNER_ATTR flow from 3.5 to BI 7.0 which has InfoSource. So i copied update rules and trasfer rules from the existing flow as both have routines. When i tried to create transformations from update rules, it propted m

  • Can not opening the file

    My program must opens file "abc.txt" which cotains float numbers (one number in aline) and stores its contents in an array of float type but it does not.it has some errors i could not fix them i am just beginner.& the open file code based on another

  • How can i run my java Application stand alone?

    Dear all, As the topic i want to run my application in a single file without showing the dos prompt or type in the command in dos prompt by the user...... How can i do that, since in my application , i have used so other package? In fact i just want

  • The program shany ip surrvilance Playback.exe version 4.0.7.3 stopped interacting with Windows and was closed

    in windows server 2008 r2 running shany ip surveillance when i want to playback from record file windows encounter error:The program shany ip surrvilance Playback.exe version 4.0.7.3 stopped interacting with Windows and was closed