Dimensional modeling using Conformed and Unconformed dims

Hi,
I have an issue in modeling the logical tables for reporting req. I have two different star schemas in logical layer. there are some reports working fine on each schema. I got new requirement where I need to use both the schemas together.
schema 1: Fact1, D1, D2, D3
schema 2: F1ct2, D1, D4
there is one conformed dimension D1 between two fact tables. but when I using attribues from D1 and measures from Fact1 and Fact2 together I'm getting below error.
Odbc driver returned an error (SQLExecDirectW)....
Error Details
Error Codes: OPR4ONWY:U9IM8TAC:OI2DL65P
State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 14026] Unable to navigate requested expression: Fact - MBS FYProject Accounts.ACCT_VALUE. Please fix the metadata consistency warnings. (HY000)
Please advise how to fix this problem?
Appreacite for your help
Thanks
Jay.
Edited by: Jay on Sep 13, 2011 11:02 AM
Edited by: Jay on Sep 14, 2011 7:14 AM

Hi,
I followed the same configuration and still getting erro shown below
nQSError: 14026] Unable to navigate requested expression: Fact - MBS FYProject Accounts.TOTAL_VACANCIES. Please fix the metadata consistency warnings. (HY000)
Dim D1 , D2 have physical joins with both the fact tables F1 and F2 and D3 joined to F1 only. I created logical fact table with LTS F1 and F2. and set logical level for F1 at details level to all dims D1, D2 and D3. and set logical level to F2 at grand total level for D2 and details level to D1 and D2. dimension hierarchies are created for D1, D2 and D3.
when I try to run a report with attributes from D1, D2, D3 and measures from Fact(F1 and F2) I'm getting above error. I'm not sure what's wrong with my configuration. if I run report with columns d1,d2 and d3 and f1 it is working fine. when I take columns from d1, d2 and f2 it is also working fine but when I combine all together is a problem.
Appreciate for your help..
Thanks
Jay...
Edited by: Jay on Sep 20, 2011 8:20 AM
Edited by: Jay on Sep 21, 2011 10:52 AM

Similar Messages

  • Cannot query using both conforming and cached query result

    TopLink doesn't allow me to both use conforming and cached query result at the same time.
    Conforming is certainly not a superset of the [cached query result] features.
    Can you confirm that it's a limitation of TopLink?
    Any know workaround to end-up with the same features as using both conforming and cached query result?
    Conforming is about seeing modifications you do in the same transaction. As a bonus, if you query for one object and specify at least the id as criteria because TopLink will have to check in memory anyway it can avoid going to the database.
    But if I do a query like "give me employees hired before now and after 30 days ago" it's about more than one objects and about finding existance so cached query result is needed to get acceptable performance in a complex application trying to avoid the same SQL generated over and over again.

    Thats where the trace just ends? It doesnt look like there's any LIKE or filtering going on (with respect to the Oracle pieces anyway), apparently MSAccess simply requested the whole table.
    What do you mean by 'hang' exactly? Are you sure it's just not taking a long time to complete? How long have you waited? How fast does it complete on the other environment?
    ODBC tracing isnt likely to help much for that. SQLNet tracing would be better to see what is going on at a lower level. Specifically, what is going on at the network level? Is the client waiting for a packet to be returned from the database?
    Is the database having a hard time processing the query, perhaps due to index/tuning issues?
    Assuming that is indeed the query that is "hung", how much data does that return?
    Are you able to reproduce the same behavior with that query and vbscript for example?
    Greg

  • Designing Dimensional Model from Reports

    I'm currently desiging a data warehouse for our project. My problem is that we do not have access to the source system. The only thing we are provided with is the data feeds that are generated by the current reporting system. This is causing me a great deal of headach b/c reports do not give you an entire picture of what the source system. Since reports only display data based on how it is developed, it is difficult to come up with an efficient design. My question is has anyone designed a dimensional model using only Reports as the source?

    Hi there,
    The short answer to your question is 'no'
    A longer answer is 'Surely it is possible to get a copy of the source schema definition even if it is not possible to get access to the source system?'
    Stuart

  • Currency Dim and Groups Dim for LegalConsolidation and IcMatching App

    Hi experts,
    I have had a look on posts about using Currency and Groups Dim, and using only Groups,
    I have developed apps with only Group Dim. Do you recommend to use Curr and Groups in the future.?
    And which dims should I include in ICmatching App with curr and groups dim?
    Regards

    Hi,
    While what the guys before have said is correct, I have a different opinion on the matter from my experience (and from the SAP starter kits since 7.5).
    I use a separate Group and Currency dimension because you may want to look at an entity's numbers in GC but without running Consolidation. For example, an entity that is only partially owned will only have GC values for its owned %, but the accountant for this entity may need to answer questions on the GC values based on 100%. With separate dimensions you can see values in GC before and after Consolidation.
    For an IC Matching application, I would have both Group and Currency dimensions. I would definitely use Group dimension because you may want to see eliminated IC balances in different groups. You should also have the Currency dimension as although only GC Interco balances are essential for matching, the LC values for each position is important as a frame of reference for the local accountant.
    Hope this helps,
    Tom.

  • Are there any (dis)advantages in building a universe on fully normalized tables instead of building dimensional model/tables and then universe on top of them?

    Hello,
    I’m hoping someone can help me with understanding advantages and disadvantages if we want to build a universe on top of a fully normalized tables, compared to using a dimensional model with star schemas.
    I’ve read some discussions here that say that it is possible to create a universe on top of normalized tables. Then, can we avoid building of dimensional tables (a data mart), and just use normalized tables? I would say that it is easier to use star schema dimensions and facts tables to build a universe, but our end users might ask “why do we have to go through building a dimensional data mart, if we can have same reports with hierarchies and drill-down functionality based on a universe built on top of our already existing normalized tables?”
    Can you point me to some established best practices regarding using normalized tables to build a universe? Any documents with some examples for this?
    Any expected difficulties during design/development phase of our universe, related to using normalized tables?
    Any expected performance degradation if we use normalized tables compared to using dimensional tables?
    Can I build universe more easily if I transform (modify) our normalized model (by using alias tables and views) to look like snowflake model?
    I’m using BOE XI 3.1, tables are in Oracle 11.2.
    Thank you

    Few Disadvantages that I usually face when building universe on Normalized tables,
    1. Performance - Read operations have to suffer because indexing strategies do not go well with table joins
    2. Derived Tables - Due to complex Queries/ Logic, most of the time, end-up creating derived tables, which doesn't use back-end table indexes and slows down the report runtime.
    3. Normalized table/ Transaction tables may not always have proper cardinality and often results in Cartesian products
    4. Normalized tables may not have tight referential integrity and may have to join more than one column and join on varchar, etc whereas, good Dimensional model datawarehouse will have proper keys/ integer joins and not always necessary to join on multiple fields
    5. Often deal with Fan and Chasm Traps
    6. Dealing with Database fields with nulls, blanks, date in numeric format, etc.,
    7. No Facts, Dimensions separated and most of the time they are in same table
    and More...
    If performance is not a matter and building Datawarehouse is a big deal, then I will start building Universe on normalized tables by having the database diagram as reference for Joins, contexts, etc
    Note: After dealing with universes based on normalized tables for few years (by creating views, complex sql queries data loading to tables and unv on these tables, derived tables), I ended up creating star-schema dimensional model (couple of months extra ETL work), users/ developers felt lot better when they have to create standard/ ad-hoc reports and they are super fast compared to previous universes.

  • I am trying to use an external 3g modem made by Huawei model e173s-6 and when I do connect it on my usb port the macbook pro just do not recognize it. I own a 2012 model with mountain lion 10.8.2 . What should I do ?

    I am trying to use an external 3g modem made by Huawei model e173s-6 and when I do connect it on my usb port the macbook pro just do not recognize it. I own a 2012 model with mountain lion 10.8.2 . What should I do ?

    I still don't know what happened to my test Aperture Library after the update to 3.4...Not only were my previews lost, but so were the original images!  Needless to say, I haven't bothered to update the main library on the Mac Pro because I don't want to deal with this level of uncertainty (even with backups in place).

  • My camera Canon G3 (i know it's old) no longer connects to iPhoto (8.1.2) on my iMac, I have checked with my friends iMac same model iPhoto version and it works fine using my cable, I would appreciate any help getting this sorted

    My camera Canon G3 (i know it's old) no longer connects to iPhoto (8.1.2) on my iMac, I have checked with my friends iMac same model iPhoto version and it works fine using my cable, I would appreciate any help getting this sorted

    As a Test:
    Hold down the option (or alt) key and launch iPhoto. From the resulting menu select 'Create Library'
    Import a few pics into this new, blank library. Is the Problem repeated there?

  • Do the Mac minis support using AirPlay and all of its functions?   I thought I saw a comment yesterday that suggested the Mac mini (particularly the late 2012 models) does not support AirPlay.  Thank you.

    Do the Mac minis support using AirPlay and all of its functions?   I thought I saw a comment yesterday that suggested the Mac mini (particularly the late 2012 models) does not support AirPlay.  Thank you.

    Hello,
    AirPlay Mirroring
    Requires a second-generation Apple TV or later. Supports the following Mac models:
    iMac (Mid 2011 or newer)
    Mac mini (Mid 2011 or newer)
    MacBook Air (Mid 2011 or newer)
    MacBook Pro (Early 2011 or newer)
    http://support.apple.com/kb/HT5444

  • How to find which model of iPhone 5 I am using? and is that GSM or CDMA?

    How to find which model of iPhone 5 I am using? A1428 or A1429 ? Is there any short code you can dial in and check the model number ?
    And which is GSM or CDMA out of this model.

    If you purchased the phone in the US with an AT&T contract then you have the locked CDMA A1428 version of the iPhone 5. 
    At the moment and if the contract is a new contact, then you won't be able to get the phone unlocked at the moment and no, you won't be able to use it in your home country.  You will also have LTE 4G compatibility issues too and your phone will only work on 3G and also you need to remember that iPhones are only covered by country specific warranty, so it is only covered in the US, so if you go home and then have a problem with your phone, you will have to come back to the US to get it repaired.
    You need to approach AT&T about whether they will unlock your phone for you as they are the ONLY ones that can do it, but the answer to your question is no and if you had wanted to use your phone at home, then you should not have purchased it in the US.  The best that you can hope for at the moment is to use it on international/roaming.
    If you purchased your phone in the US, then it will be a CDMA phone.  You don't have a GSM phone, but GSM phones are unlocked anyway and these are the ones that are selling in Canada and the UK.
    If you have your phone on an AT&T contract then ultimately, AT&T are the only ones who can help you with any questions regarding unlocking and they are the only ones who can unlock it for you, noone else.

  • How to setup and configure NI OPC Server to OMRON PLC (CJ2H model) using FINS Ethernet communication protocol?

    Hi,
    Could someone help me in the setup and configuration of NI OPC server to OMRON PLC (CJ2H model) using FINS Ethernet communication protocol? I already read some materials available in the web but it uses a different PLC and communication protocol. The OPC server is installed in my laptop and I'm using Windows 8 as my OS. I used a cross over RJ45 LAN cable and connected to the ETN21 module of the PLC. I was able to confirmed the communication of my laptop and PLC by pinging the ip address of the ETN21. I also tried to configure the OPC server based on the documents I just read online. I created a channel, device and tag to the OPC server. By the way, I just need to read the data from the tag I created. Can someone also tell me how to monitor the data on the tag? I hope that you could help me. Thank you in advance. 

    Good day,
    Try following the instructions on the driver’s help, take a look at the attach picture. It should be a good starting point. After the PLC is configured you can use this tutorial (Connect LabVIEW to Any PLC Using OPC » http://www.ni.com/white-paper/7450/en/) for a guide on how to monitor the tags.
    I hope this helps
    Alejandro | Academic Program Engineer | National Instruments
    Attachments:
    OPC Server Help.JPG ‏237 KB

  • IPad 1 3G wifi Model using original operating software and video streaming

    Anyone experience the following ?  My iPad 1 , 3G wifi model , using  original operating system software ... When using 3G , AT&T plan, when streaming video clips from sites like YouTube, tmz, xvideos, etc, although slow, all streams came back crystal clear, phenomenally clear , for some reason, every time , no matter if the clip was hd or not .  This doesn't happen when I used wifi.      Now that I have downloaded iOS 5.0.1 to my iPad, it's not the same.  Any help advice or comments, or anyone else experience this before? 

    jrhee272 wrote:
    Asatoran wrote:
    iOS5 is a bit slower on the iPad1.  And from what I'm reading, a tad bit slow also on the iPhone 3GS and 4.  All of these use the A4 processor, versus the A5 in the iPad2 and iPhone 4S.  So that may account for some of the slowness of the video playback....
    What would account for the crystal clear streaming videos when I was operating original iPad software using AT&T wireless plan?  All streaming video came back crystal clear, even though they weren't originally hd streams. 
    Like I said, iOS5 is slower than iOS4 on the iPad1.  Slower so can't decode the frames quickly enough before the next frame has to be displayed.
    Ane like I inferred, cellular data is inconsistent.  More users in your area will cause your speed to decrease.  You can't really guarantee enough bandwidth on a 3G connection in many areas.  Wi-Fi would almost certainly use a faster connection than 3G, thus the better video playback, regardless of iOS4 or iOS5.
    Bottom line: 3G in your area is probably just not good enough anymore.  iOS5 just made it worse for your already marginal situation.

  • I have bought a New Zeeland iPhone 5 model MD297X/A and I'm wondering if I will be able to use the 4g network in Sweden?

    I have an iPhone 5 that I bought down in Australia. I can read from the model number that its a New Zeeland registered model! I'm going back to Sweden and I want to know if I'm going to be able to use it and if I can use the Swedish 4g carriers? It's model number: MD297X/A

    You can use the phone in Sweden if it is unlocked, but you won't get a 4G signal because all the frequencies vary greatly between countries, especially European countries and Australia/New Zealand, so it won't be compatible and they can't cover all frequencies in one phone, which is why different models are sold in different countries.
    If you want to use your phone in Sweden on the Swedish 4G networks, then you'll have to purchase a new phone in Sweden that will work with the 4G network.  Your other phone will work, but NOT on 4G.

  • Dimensional Modeling on SAP Hana

    Hello,
    I am a newbie on SAP hana, so need help from experts / gurus.
    Imagine the following Scenario ---> Oracle ERP is the source System and we want to build a Analytic Solution on Hana. Also assume that tables can be replicated into SAP Hana
    Questions
    1)For any kind of reporting / Analytics? Should we build a traditional Datawarehouse on Hana with all the features found in a star schema like surrogate keys, conformed facts and dimensions etc? or  use only attribute views / analytical views provided as default and build datawarehouse model on top of this? In the second case, do we still need surrogate keys or how can we merge slowly changing dimensions?
    2)Where can be push the transformation logic? Into ETL tool or use Hana provided sql script? and also where do we do this transformation, outiside hana where we do all transformations and load only data into normolized format? or within hana using scripts?
    3) In any case is dimensional modeling really required on Hana? if not how can we map business process and do analysis on them? is it by means of only the views provided by Hana?
    Your help is much appreciated.
    Thanks!

    Hi Rajiv,
    I'm no expert either and am just beginning to learn HANA, but here are my thoughts (community, please correct me if I am mistaken.):
    1. HANA leverages existing data modeling theories, such as the entity relationship model. This model is simply a way of organizing tables in a way that makes sense for reporting. For example, the star schema, optimized for analytical reporting. The HANA view is similar to a universe, it allows the end user to connect to it from their BI tool. In the background, the modeler still needs to pull in tables, create joins, leverage primary and foreign keys, etc. So in short, traditional modeling practices and theories still come into play.
    2.  If extracting and transforming data using SAP's Data Services, the transformation of data will happen in HANA, using the Data Quality Management GUI.
    Just as a side note, HANA has 3 methods of extracting
    data:
    Data Services
    SAP Replication
    Server
    SAP Landscape
    Transformation
    3. Yes, dimensional modeling is still required. At the end of the day, HANA is still a database. We
    haven’t changed the theory of data modeling, it still needs to happen, our approach is just different.
    Hope that helps.
    Kyle

  • How to include Abbreviations in Dimensional Model

    Dear All,
    The question is basically related to modelling but as I am using OWB for this, I am posing the question here.
    I have to migrate a HR database from relational model to dimensional model. I have the following situation.
    There is a table which has two columns i.e. Country_cod, and country_name like UK- United Kingdom, US - United States. Then in the database where ever it is needed to use information about the country, the country_cod is used. I want to migrate the database to dimensional model, This table does not relate to any other table directly, like a foreign key relationship.
    How should I deal with this situation?
    Comment please....
    regards

    You can put your Javascript code in the Onmouseover property of the column.
    I don't have the Javascript code for changing the color but for example the following will pop up an alert
    <af:outputText value="#{row.DepartmentId}" onmouseover="alert(\'Unload event fired!\')">

  • What r the steps that involve while v r using DBCONNECT AND UDCONNEC

    Hi All,
    What r the steps that involve while v r using DBCONNECT AND UDCONNECT AND IN WHICH DIFFERENT SCENARIOS V USE THEM.
    REGARDS
    KK

    Hi,
    To Configure UD Connect on the J2EE Server for JDBC Access to External Databases;
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/58f4db47-0501-0010-a2bf-ff01b150fdff
    To create Driver on JDBC side
    http://help.sap.com/saphelp_nw70/helpdata/en/b0/6e62f30cbe9e44977c78dbdc7a6b27/frameset.htm
    Configuring the BI Java Connectors
    http://help.sap.com/saphelp_nw70/helpdata/en/43/e35b3315bb2d57e10000000a422035/frameset.htm
    Then to extract data use this...
    Before you can transfer data from UD Connect sources into BI, you have to generate the metadata (the information about the source object and source object elements) in BI, in the form of a DataSource with a function module for extraction.
    If your dataflow is modeled with objects based on the old concept (InfoSource 3.x, transfer rules 3.x, update rules 3.x), you can generate a DataSource 3.x for transferring data into BI from a database source system.
    Prerequisites
    You have modeled the InfoObjects that you want to use in the InfoSource and the data target or InfoProvider according to the UD Connect source object elements.
    Note the following background information:
    Using InfoObjects with UD Connect
    Data Types and Their Conversion
    Using the SAP Namespace for Generated Objects
    Procedure
    You are in the Modeling InfoSource tree of the Administrator Workbench. First create in InfoSource and activate the communication structure. Then generate the generic DataSource using the wizard in the InfoSource maintenance.
    1. Choose InfoSources ® Your Application Component ® Context Menu (right mouse click) ® Create InfoSource.
    2. Select the type of the InfoSource.
    3. Under InfoSource, enter the technical name of the InfoSource, and assign a description and choose Continue.
    The system creates an InfoSource and shows it in the InfoSource tree under your application components.
    4. In the context menu of the InfoSource, choose Change .
    The communication structure maintenance screen appears.
    5. Using the previously modeled InfoObjects, create the communication structure (see Communication Structure).
    6. Save and activate your communication structure
    7. In the dialog box that follows you will be asked whether you wish to activate the dependent transfer program. Choose No.
    8. In the InfoSource menu, choose Extras ® Create BW DataSource with UD Connect.
    You get to the dialog box where you can assign a UD Connect source object to a DataSource and then generate the DataSource with the extractor. The fields for the DataSource are already being displayed in the table on the left side of the screen. The fields have the same name as the InfoObjects that you have used in the InfoSource.
    9. Select the RFC Destination for the J2EE Engine.
    Make sure that the local server is running. If the local server is running and you cannot open the table for RFC destinations, start the local server again.
    10. Choose the UD Connect Source in which the data that you wish to access is located.
    All available sources that are connected to the J2EE Engine are provided for selection in the input help. n instances are available per adapter.
    11. Select the UD Connect Source Object.
    All available source objects that are available in the selected UD Connect source are provided for selection in the input help. Source objects can be multi-dimensional storage or relational objects.
    The system generates the name of the DataSource in the namespace 6B.
    The DataSource is then replicated in BI.
    The myself system is assigned to the InfoSource as source system as well as the DataSource.
    The system generates a proposal for the transfer rules.
    Since the DDIC structure and the function module are located in the SAP namespace, the following entries can be requested during generation:
    Developer and object key or
    Developer key or
    Object key and
    Transport request
    If you do not make the requested entries, the generated infrastructure is not usable.
    14. Change or complete the transfer rules as needed. For example, if a source object element is not assigned to a unit InfoObject, you can determine a constant for the unit, such as EUR for 0LOC_CURRCY (local currency).
    15. Save and activate your transfer rules.
    Recognizing Manual Entries
    You can enter and change the RFC Destination, UD Connect Source and UD Connect Source Object manually. In order to validate these entries and the ones that depend on them, choose Recognize Manual Entries. For example, if you change the selected RFC destination, all of the field contents (US Connect Source, UD Connect Source Object, list of the source object elements) are invalid. If you choose Recognize Manual Entries, the dependent field contents are initialized and have to be maintained again.
    Result
    You have created the InfoSource and DataSource for data transfer with UD Connect. In the DataSource overview for the myself system, you can display the DataSource under the application components Non-Assigned Nodes.
    When modeling InfoObjects in BI, note that the InfoObjects have to correspond to the source object elements with regard to the type description and length description. For more information about data type compatibility, see Data Types and Their Conversion.
    The following restrictions apply when using InfoObjects:
    Alpha conversion is not supported
    The use of conversion routines is not supported
    Upper and lower case must be enabled
    These InfoObject settings are checked when they are generated.
    Based on the large number of possible UD Connect sources, the most diverse data types are possible in the output system. For this reason, a compatibility check is made at the time of generation of the UD Connect DataSource that is based on the type information supplied by the source systems. This attempts to decrease the probability of errors during the extraction process.
    (Refernce:SAP Notes)
    Hope this helps,
    regards
    CSM Reddy

Maybe you are looking for

  • Use of Prepared Statement in adf

    Hi Experts, I am confused with the Use of prepared statement in adf. My use case is , I have to update a table from every page in my application under certain conditions. My question is , whether I have to create the VO iterator binding in every page

  • Win 8.1 stutter issue with Flash 15.0.0.189

    Greetings - I am having stuttering issues with 2 brand new 8.1 computers running flash. I've done a fair amount of troubleshooting and I believe it's an issue with Win 8.1 and Flash 15.0.0.189 (I don't know about previous versions. Running this versi

  • Template for a conference room bookings

    Hi - we are planning a conference and we need to keep track of room bookings and costs. is there such a template we could use?

  • EP6SP2 Portal patch 4prob: java developement - documentation not running

    hi, Before applying EP6SP2 portal patch 4 everything was working fine ... but after applying portal patch 4 <b>java developement -> Documentation and examples tabs</b>are not working ... It is giving <b>Portal runtime error ....... page could not cre

  • Material Characteristics

    Hi I have already posted this doubt. In MM03 if i go to Material Classification, i get a list of description based on the Class. Now as an ABAPer i need to get the Description. I dont know the Tables. I used the FM 'QMSP_MATERIAL_BATCH_CLASS_READ' to