Data Model: Keeping models in sync

Morning,
I'm not clear what exactly i need to do to keep models in sync?
Here is the situation:
1. Original DB (schema) doesn't have a logical model avaialble.
2. I reverse eng the schema and proceeded working on the logical model. My changes to the logical model include creation of the new tables and adding a few FKs to the existing tables.
3. Now i need to incorporate all changes that have been done to the physical DB since i last did a rev. eng.
I read the manual, but it s not clear what and how i need to proceed.
For example:
i can rev eng (file -> import -> Data dict) from the same db schema. This creates a new relational model. Engineering into logical model produces copies of the existing tables.
I tried other options - create 2 separate model ( files ) and go through the merge. But it looks convoluted and i end up with another model (file).
I remember in old times it was pretty easy in ERWin to incorporate changes .
Please let me know the "right way" .
thank you
Andrew

Sue, thank you. Your answer is helpful.
A note for the demo: I found it important to specify the "Database Types" -> "Database Type" during the "merge" operation. The results of the merge will be different if you leave the option empty.
Second,
Can you please explain the meaning of "yellow exclamation" marks and how "selection" of the tables/views etc works in the "tree"? I understand that "green +" sign is for "new" objects.
I'm 99% certain that "yellow !" is for changes. But mechanism of assigning the "selection" mark on tables in the "tree" is still a puzzle.
Another "Gray" area:
I merged 2 relational models - original and update. Let's call them original -v1, latest obtained by rev. eng of the DB -v2 and the merged version of v1 and v2 => v3.
So if i try to merge v3 and v2 ( yes, v2) i'm getting a bunch of yellow signs and marked tables again.
Next:
I noticed that even if "mark" is set on the level of the "table" it doesn't mean that all columns and FKs ( in the table ) are selected. The only way to insure that they are (selected) is to un-mark the table and mark it again.
Why?
thank you
Andrew

Similar Messages

  • Data Modeler bug? - it keeps reporting missing physical model files

    Hi,
    I'm having a rather serious problem with the latest data modeler, Oracle SQL Developer Data Modeler 3.1.4.710. It keeps losing physical model files for some reason.
    I started to notice that it was not generating all the tables in my small design. After hours of struggle I found multiple error messages in the log file:
    2012-12-18 10:51:20,126 [Thread-71] ERROR XMLTransformationManager - Unable to load object from XML: C:\cygwin\home\Brian\codebase\models\data\Design\rel\395D160B-C7224B8EAD05\phys\D9582E4E-2ED963CB9D32\View\seg_0\A555DB9B-8864-9C0B-DCA8-BAC5231F55C6.xml
    java.io.FileNotFoundException: C:\cygwin\home\Brian\codebase\models\data\Design\rel\395D160B-C7224B8EAD05\phys\D9582E4E-2ED963CB9D32\View\seg_0\A555DB9B-8864-9C0B-DCA8-BAC5231F55C6.xml (The system cannot find the path specified)
         at java.io.FileInputStream.open(Native Method)
         at java.io.FileInputStream.<init>(FileInputStream.java:138)
         at oracle.dbtools.crest.model.metadata.XMLToObjectTransformer.transformToObject(Unknown Source)
         at oracle.dbtools.crest.model.metadata.XMLToObjectTransformer.transformToObject(Unknown Source)
         at oracle.dbtools.crest.model.metadata.XMLTransformationManager.transformFromXMLToObject(Unknown Source)
         at oracle.dbtools.crest.model.metadata.XMLTransformationManager.openPhysicalModel(Unknown Source)
         at oracle.dbtools.crest.model.metadata.XMLTransformationManager.openPhysicalModel(Unknown Source)
         at oracle.dbtools.crest.swingui.browser.BrowserPanel$88.run(Unknown Source)
    These messages appeared every time I opened my physical model (Oracle 10g).
    After deleting and recreating the physical model the problem went away for a while. But after closing and reopening the design I started to see the errors again.
    Has anyone seen something similar?
    Regards,
    Brian

    Thanks Philip!
    The workaround does the trick for the tables, primary keys and unique keys.
    I have just observed that table level constraints show a slightly different behaviour.
    Names show up like TCC5, TCC6, etc.
    Seems the same problem as previously in SQL*Modeler forgets check constraint names
    Is there a workaround for that problem too ?
    I already noticed that there is no parameter NameChanged for a table constraint.
    Enrico

  • Tabular data model: Query keeps timing out when attempting to Edit Table Properties

    Tabular data model (SSDT)
    Problem: I have a table in tabular data model using a SQL Query for a data source. The query in question requires about 3 minutes to regenerate. When I open Edit Table Properties for this data source the query times out and I get an error (see below): "
    Failed to retrieve data from udvTrainJobReportsData. Reason: Query timeout expired"
    This seems to happen anytime I use a query that takes longer than a couple of minutes to regenerate. Anyone have an idea on how to get around this. Is there a timeout setting somewhere in tabular data model that can be increased?
    Thanks...

    Hi ManikantM,
    According to your description, you query keeps time out when edit table properties. Right?
    In this scenario, this error is thrown when connection or query execution exceeds the time out value. Please try to import this table and then increase the connection time out seconds.
    We can increase to ExternalCommandTimeout in Analysis Server Properties. Please refer to link below:
    http://aniruddhathengadi.blogspot.in/2012/07/ole-db-error-ole-db-or-odbc-error-query.html
    Please also refer to a similar thread below:
    https://social.technet.microsoft.com/Forums/office/en-US/3f83a26b-71c6-462e-8b90-2ce2ce0b9465/powerpivots-2010-query-keeps-timing-out-when-attempting-to-edit-table-properties?forum=excel
    Best Regards,
    Simon Hou
    TechNet Community Support

  • SQL DM - Copy/Export data models - bug?

    SQL DM - Copy/Export data models - bug?
    I am using SQL DM 3.0.0.665. I need your thoughts on following.
    We used saved as/export of the SQL DM copy the model into another location. We kept our models in harddrive as folder. We did not store it in database.
    We opened the copied model from target model. We found all the links to defined domain was lost. We find the data type shows as Varchar2 instead of Varchar2(20)etc. It creates error on sql.
    Why are loosing the domain definition/linkages during export/save as option?
    Is it a bug?
    Did anyone face such issues?
    What is the workaround?
    Thanks in helping us out.

    Hello,
    Thanks for your reply. I followed the steps. I did copy the folders/files as Kent suggested.
    As we installed the SQL developer on new desktop, it showed the domain definitions as unknown and count as 1 in Design Tree --> domain node. We copied all the xml files under C:\datamodeler\datamodeler\types folder of Desktop A to Desktop B. It brought all the domain definitions of Design in Desktop A.
    Why does it not keeping this domain definitions as part of Design folders?
    If we copy or save as the design , then domain also cane be saved as part of design. It will be difficult to make a sync of datamodels among multiple desktops.
    How do we establish a shareable/sync data models using SQL data Modeler on such situations?
    Thanks in helping.

  • Some observations on my first use of Data Modeler Beta

    First of all, I can see this tool has a lot of promise.
    I hope Oracle keeps at it, it could turn into a real winner if all the features I see being worked on mature.
    Thanks!
    Here are a few observations on things that I found non-obvious or tedious to do.
    1. When designing an entity, I want to give it a name, a definition, attributes and keys. I want that process to be quick and require the minimum amount of mouse-clicking/navigation fiddling as possible. The current way of defining the attribute's datatype and size is painfully slow. I have to click to get a pop-up. Then I have to click to choose from a set of categories. Then I have to click on a dropdown list. If I try to use the down-arrow on the dropdown list, it works, but not if I go past the one I want. The up-arrow won't take me backwards in the list, so more clicking. It's just a nasty, slow interface to do a simple task that I have to do a thousand times in a data model. If I need to change the size of something, back I go thru the entire process all over again.
    That makes it doubly slow to work in the most natural way, which is to list the attributes and the datatypes, then come back and refine the sizes once the model is maturing and relatively stable.
    2. Adding an additional attribute requires a mouse click instead of a down-arrow. That means I have to take my hands off the keyboard to add a new attribute. Maybe there is some short-cut key that does that, but I have better things to do than memorize non-standard keyboard mappings. Make the down/up arrows works as they should.
    3. Adding the comment that describes the attributes is not quite as slow, but still requires more keystrokes/mouse movements than it should. It's hard enough to get developers to document their attributes, don't discourage them.
    4. I can't see the list of attributes, data types, sizes, key/mandatory settings, and the comment at one time while editing the entity. Makes it harder to grasp what all the attributes mean at a glance, which slows down the modeling and the comprehension of an existing model.
    5. All the entities I created had primary keys with columns in them. But when I tried to have it build a physical model, it complained that some relationships had no columns in them. For the life of me, I couldn't figure out how to fix that. Never had that problem in any other case tool.
    6. Getting it to generate DDL was awkward to find. Make it something obvious, like a button on the toolbar that says "Generate DDL".
    7. Apostrophes in the Comments in RDBMS are not escaped, so the generated DDL won't run.
    8. For the ease of use/speed of use testing on high-volume key tasks, make the developers do the task 1000 times in a row. Make them use long names that require typing, not table A with columns c1, c2 and c3. Long before they get to iteration 1000 they will have many ideas on how to make that task easier and faster to do.
    9. Make developers use names of things that are the maximum length allowed. For example, for a table name in oracle, the max length of the name is 30 characters. The name of one testing table should be AMMMMMMMMMMMMMMMMMMMMMMMMMMMMZ. That's a capital A followed by 28 capital M's and a capital Z. For numbers, use the pattern 1555559. If the developers can't see the A and Z or 1 and 9 in the display area for the name in the default layout for the window, they did the display layout wrong. For places where the text can be really long, choose a "supported visible length" for that field and enter data in the pattern AMMMMMMMMMMMMMMMQMMMMMMMMZ, where Q is placed at the supported visible length. if the A and Q don't show, the layout is wrong.
    10. SQL Developer has quite a few truly gooberish UI interaction designs and I can see some of that carrying over to the Data Modeler tool. I really recommend getting a windows UI expert to design the ui interface, not a java expert. I've seen a lot of very productive windows user interfaces and extremely few java interfaces suitable for high-speed data entry. Give the UI expert the authority to tell the java programmers "I don't want to hear about java coding internals - make the user interface perform this way." I think the technical limitations in java UIs are much less than the mindset limitations I've seen in all to many programmers. That, and making the developers use their code 1000 times in a row to perform key tasks will cause the UI to get streamlined considerably.
    Thanks, and keep up the good work!

    Dear David,
    Again thank you for your valuable and highly appreciated feedback. Find included a more elaborated answer to your observations:
    *1. When designing an entity, I want to give it a name, a definition, attributes and keys.*
    In the new Early Adopter Release, a "SQL developer like" property window has been added for Entities and Tables. For each attribute/column you will have the ability to add name, datatype, Primary Key, Mandatory and comment from one and the same screen
    *2. Adding an additional attribute requires a mouse click instead of a down-arrow.*
    An enhancement request has been created
    *3. Adding the comment that describes the attributes is not quite as slow, but still requires more keystrokes/mouse movements than it should.*
    In the new Early Adopter Release, a "SQL developer like" property window has been added for Entities and Tables. For each attribute/column you will have the ability to add name, datatype, PK, M and comment from one and the same screen
    *4. I can't see the list of attributes, data types, sizes, key/mandatory settings, and the comment at one time while editing the entity. Makes it harder to grasp what all the attributes mean at a glance, which slows down the modeling and the comprehension of an existing model.*
    See former answers. For meaning of attributes you can also use the Glossary and Naming Standardization facilities: see Tools Option menu, Glossary and General Options for naming standards
    *5. All the entities I created had primary keys with columns in them. But when I tried to have it build a physical model, it complained that some relationships had no columns in them. For the life of me, I couldn't figure out how to fix that. Never had that problem in any other case tool.*
    A Bug report has been created. Issue will most probably be solved in the nexr Early Adopter release
    *6. Getting it to generate DDL was awkward to find. Make it something obvious, like a button on the toolbar that says "Generate DDL".*
    An enhancement request has been created.
    *7. Apostrophes in the Comments in RDBMS are not escaped, so the generated DDL won't run.*
    A bug report has been created
    *8. For the ease of use/speed of use testing on high-volume key tasks, make the developers do the task 1000 times in a row. Make them use long names that require typing, not table A with columns c1, c2 and c3. Long before they get to iteration 1000 they will have many ideas on how to make that task easier and faster to do.*
    I aplogize, but I don't understand clearly what you want to say with the use/speed of use here.
    *9. Make developers use names of things that are the maximum length allowed.*
    Our relational model is for use for not just Oracle, but also DB2, SQL Server and in the future maybe other database systems. Whicjh means that we can't taylor it to just one of these database systems. However you can set maxinum name lenghts by clicking right on the diagram and select Model Properties and here you can set naming Options. Here you can also use the Glossary and Naming Standardization facilities: see Tools Option menu, Glossary and General Options for naming standards
    *10. SQL Developer has quite a few truly gooberish UI interaction designs and I can see some of that carrying over to the Data Modeler tool.*
    Fully agree. As you will see in our next Early Adopter release we have started to use SQL Developer like UI objects.
    Edited by: René De Vleeschauwer on 17-nov-2008 1:58

  • APEX application development with an existing data model

    Dear all,
    We - as a company - are trying to build an application in ApEx with an existing data model. The idea is that the data model that has all sorts of TAPIs and business rules defined is going to be re-used in an ApEx environment. I am actually wondering whether this is possible, wise, feasible. When I am building some simple pages on one specific table, which has approx. 35 fields, sometimes the triggers around that table fail. Before I am actually going to dive in to try and sort these problems I am wondering whether this use of a data model in an ApEx environment is wise.
    I mean, when you build an ApEx application there are usually no triggers and TAPIs available. My logic is that the ApEx application can look after validation and stuff.
    Does anybody experienced have anything smart and useful to say about this. Any feedback is appreciated.
    Kind regards,
    -victorbax
    -leiderdorp, the netherlands

    Hey vik,
    At my company we rarely use the standard APEX wizards because of multi-linguility (on data level) and error management issues.
    We always create an API in PL/SQL and call the procedures for APEX making APEX the 'view' layer in a kind of model-view-controller architecture.
    The main thing to keep in mind is maintainability, if you're working with an existing datamodel with an API that works it's probably best to just created forms and reports through APEX and then calling the appropriate pl/sql (packaged) procedure manually. We believe that this kind of architecture has the downside of being a little more work initially but it's much easier to maintain large projects through a self written API and take care of issues like error handling, logging and multi-linguility through a self written API.
    Kind Regards,
    Geert Guldentops

  • Data model: single quote in default parameter value

    Hello,
    when I assign a default value to a parameter in my Data model, which includes a single quote (i.e. "It's a default value"), then the single quote is escaped with a backslash (i.e. "It\'s a default value") when I open the report associated with the data model. How can I get rid of the escape character when opening a report?
    Regards,
    Jure

    Hi Yogini,
    Thanks for your reply.
    When the customer name is selected, discoverer reports automatically enclose the customer names in single quotes. like 'ABC LTD', 'ABC's INC',
    I know how to change this single quote format. Also we don't want user to change the parameter selection criteria once selected.
    Also related to double quotes isn't it doubles are treated as column names in the SQL query and can not be used for any other purpose.
    Apologies, I haven't tried this but I think it will not work. Have you tried this with discoverer viewer.
    As mentioned intrusting things is when a customer name with single quote is selected in discoverer Plus it automatically formats the filter criteria to look like ['ABC LTD', 'ABC''s INC'] but this thing doesn't happen in viewer.
    I think this is a bug in discoverer viewer and think there must be some patch available to correct this problem.
    Sorry, don't have a solution yet but will certainly update the thread once I get the solution.
    Also please keep looking for a solution and let me know as well.
    Thanks a lot in advance.
    Best Regards,
    Manish

  • Building BI Publisher report with "BI Answers" data model

    I've installed OBIEE 10.1.3.4.1 and try to create a BI Publisher report using BI Answers as data model. The catalog list of Answers comes up empty, and in the log files I find following error:
    - Exception:
    AxisFault
    faultCode: {http://schemas.xmlsoap.org/soap/envelope/}Server.userException
    faultSubcode:
    faultString: java.lang.IllegalArgumentException: The char '0x14' after '��u���3' is not a valid XML character.
    faultActor:
    faultNode:
    faultDetail:
    {http://xml.apache.org/axis/}stackTrace:java.lang.IllegalArgumentException: The char '0x14' after '��u���3' is not a valid XML character.
    at org.apache.axis.components.encoding.UTF8Encoder.writeEncoded(UTF8Encoder.java:81)
    at org.apache.axis.encoding.SerializationContext.writeSafeString(SerializationContext.java:1289)
    at org.apache.axis.encoding.ser.SimpleSerializer.serialize(SimpleSerializer.java:104)
    at org.apache.axis.encoding.SerializationContext.serializeActual(SerializationContext.java:1504)
    at org.apache.axis.encoding.SerializationContext.serialize(SerializationContext.java:980)
    Does anyone have a clue?
    Thanks,
    Marc

    option1:
    The query should be same @ detail level, only the template has to be different for summary and details.
    @runtime, user can choose to see the detail/summary
    Disadvantage, if the data is huge,
    advantage , only one report.
    option2:
    create two separate reports summary and details
    and create diff data and diff layout and keep it as different report
    Advantage, query will perform based on the user run report, summary/detail, so that you can write efficient query.
    Dis advantage , two reports query/template to be maintained.

  • Can SAP leverage MM and classification (CL-system) data-model?

    Dear all,
    At the moment I’m working for a telecom company, setting up the Functional Management department activities to manage 2nd line support calls, and general overall application functional support.
    One of the issues is simplifying the work of monitoring Item (products, articles etc.) Master Data. Below you can see the data model where Items, their classes, the characteristics of classes and prices are stored. 
    The Legacy system has the following data-model (as ERD, <a href="http://sapparking2007.freehostia.com/LegacyModel.JPG">LegacyModel</a> )
    Main Product (i.e. Nokia 6110)
         - has zero, one or more Characteristics (i.e. 220v cord, 3v battery etc.)      
         - has zero, one or more Variants (i.e. Nokia 6110i,  Black, Red etc.)
         A Variant  (Nokia 6110i, Black)
              - has zero, one or more Characteristics (i.e. 200h stand by time etc.)      
              - has zero, one or more Deliveries (i.e. we sell “a package”)
              A Delivery (certain combination of a package : Nokia 6110i, Red, Prepaid, 1 year)
                   - Has zero, one or more Characteristics (i.e. prepaid simcard, free service etc.)
                   - Has zero, one or more Delivery Prices (i.e. prices)
                   A Delivery Price
                        - Has zero, one or more Characteristics (i.e. 1 month warenty)
                        - Has zero, one or more Prices (i.e. offer price, sales, periods)
    Well, almost perfectly normalised data-model.
    The implementation in SAP is as follows:
    Main Products , Variant and Deliveries are all stored in table MARA.
    Product is Artikeltype Z01A,
    Variant is type KMAT,
    Deliveries is type HAVA and DIEN.
    See image:
    <a href="http://sapparking2007.freehostia.com/Model01.JPG">Current Data Model</a>
    The Characteristics are stored in table CABN etc…
    The prices are stored in the SD table KONP .. (price condition). etc.
    The MARA table is connected to Classes table (KSSK) via the INOB table.
    Classes are connected to Classes with in de KSSK table.
    The filed that stores Class in the KSSK and the field that stores the Class below are 2 different type of fields (18C , and 50CN)  and they would simply wont join in a SAP query.
    The filed that stores Item to Class in the INOB and the field that stores the Class in the KSSK are 2 different type of fields (18C , and 50CN)  and they would also wont join in a SAP query.
    Therefore we have created extra Z. “Tables” (Types actually) that does the join for us.
    Functional Management personnel – are not ABAP and SQL people. So we have to keep it simple.
    After a training session – it proved to be still too complicated. After a Change Request, to create the product structure in a separated, single table, has resulted in a batch job, that create the file periodically, but not all of it. And – at the end, you never know if your data is up-to-date.
    To me it seems logical – to add a (derived) field to the KSSK and the INOB to accumulate the underlying class with the same attributes as the MARA and KSSK, so it can be properly joined. But here there is too much resistant in conjunction with SAP releases etc. AND it seem to be impossible (not sure) to define a “type table” with a derived field.
    To let programmers produce more ABAP is not a clean option either, and yes I realise one can create a sap query on a single table, and add a long SQL statement on Record level.
    So the question:
    -     Is this implementation “the best” possible in SAP?
    -     How to overcome the “unjoinable” friendly database?
    In other query tools it is possible to create “tables” as output (on the fly), where the tables can be used as input tables in a next query, and can be joined further.
    -     Is this somehow possible in SAP?
    I wonder if anyone can shade some light on this approach. We have 2 FTE to monitor 2.5 million products combinations, and the propagation of this data to over 30 sales channels – daily.    
    Thank you in advance for taking the time!
    Kind Regards,
    Doron.

    Ashish:
    Please search the space threads before posting.  There are several posts where this is already covered.
    regards,
    bill.

  • SQL Developer - problems with the Data Modeller part

    Hi SQL Developer users,
    I can open the Data Modeler pane, but I can't insert any object into the logical model or into the relational model. Any ideas about what I'm doing wrong? Do I need specific privileges to the database, or do I need to install SQL Developer in a different way?
    Thanks,
    Csisz

    Hi Csisz,
    as Ivaylo wrote - it's Data Modeler Viewer into SQL Developer.
    I can't insert any object into the logical model or into the relational modelYou can drag tables from SQL Developer browser and drop them on relational model diagram. If you keep Ctrl key pressed then all child tables also will be imported and shown on diagram.
    Philip

  • Best practice on extending the SIEBEL data model

    Can anyone point me to a reference document or provide from their experience a simple best practice on extending the SIEBEL data model for business unique data? Basically I am looking for some simple rules - based on either use case characteristics (need to sort and filter by, need to update frequently, ...) or data characteristics (transient, changes frequently, ...) to tell me if I should extend the tables, leverage the 'x' tables, or do something else.
    Preferably they would be prescriptive and tell me the limits of the different options from a use perspective.
    Thanks

    Accepting the given that Siebel's vanilla data model will always work best, here are some things to keep in mind if you need to add something to meet a process that the business is unwilling to adapt:
    1) Avoid re-using existing business component fields and table columns that you don't need for their original purpose. This is a dangerous practice that is likely to haunt you at upgrade time, or (worse yet) might be linked to some mysterious out-of-the-box automation that you don't know about because it is hidden in class-specific user properties.
    2) Be aware that X tables add a join to your queries, so if you are mapping one business component field to ATTRIB_01 and adding it to your list applets, you are potentially putting an unnecessary load on your database. X tables are best used for fields that are going to be displayed in only one or two places, so the join would not normally be included in your queries.
    3) Always use a prefix (usually X_ ) to denote extension columns when you do create them.
    4) Don't forget to map EIM extensions to the extension columns you create. You do not want to have to go through a schema change and release cycle just because the business wants you to import some data to your extension column.
    5) Consider whether you need a conversion to populate the new column in existing database records, especially if you are configuring a default value in your extension column.
    6) During upgrades, take the time to re-evalute your need for the extension column, taking into account the inevitable enhancements to the vanilla data model. For example, you may find, as we did, that the new version of the S_ADDR_ORG table had an ADDR_LINE_3 column, and our X_ADDR_ADDR3 column was no longer necessary. (Of course, re-configuring all your business components to use the new vanilla column can also be quite an ordeal.)
    Good luck!
    Jim

  • Looking for a Good White Paper on Oracle Data Modeling

    Hopefully this is the proper area in which to ask this questions - been looking around these boards for a while but can't find what I'm looking for. I'd like to get a good modeling overview for topics such as:
    -- pros and cons of normalization, e.g., performance considerations when deciding to normalize some data (and incur the cost of joins) vs keeping it denormalized (and having simpler queries)
    -- overall performance considerations
    -- warehouse vs mart considerations, esp as they relate to front end tools that will sit on the data mart
    A general data modeling white paper or PDF would be very helpful to get me started.
    Thanks all!
    Mike

    You don't need a White Paper you need a college course or a book.
    The book I have on my shelf, don't know if it is still in print, is CASE*METHOD Entity Relationship Modelling by Richard Barker
    Addison-Wesley Publishing Company
    ISBN 0-201-41696-4
    Normalization isn't about pros and cons. For performance denormalize. For data integrity normalize. Most systems end up somewhere between 3N and 4N or with what is referred to as Boyce-Codd NF. Google is your friend here but buy the book or take the course. This is not a subject to be read lightly in a dozen or so pages.

  • BI Publisher Report with Multiple data models

    I have an Oracle Report with two different data model(querys) and Layouts for Summary and Details. When the user Choses Summary it execute the summary data model query and display the result in summary Repeating frame Layout and when user choses Details it execute the other query and uses the other layout.
    Now I am creating the same report in Stand alone BI Publisher. As a company standard we are using Data Template for Data model. How can I accomplish the same feature I mentioned above in BI Publisher? Meaning How can I control and execute the querys in BIP based on a parameter "Summary" or "Details" ?
    Can Someone help
    George

    option1:
    The query should be same @ detail level, only the template has to be different for summary and details.
    @runtime, user can choose to see the detail/summary
    Disadvantage, if the data is huge,
    advantage , only one report.
    option2:
    create two separate reports summary and details
    and create diff data and diff layout and keep it as different report
    Advantage, query will perform based on the user run report, summary/detail, so that you can write efficient query.
    Dis advantage , two reports query/template to be maintained.

  • Can't create connection with Data Modeler 2.0.0 570

    I have installed SQL Developer Data Modeler on my Windows Vista Home Premium system and am unable to create a connection to import a Data Dictionary. When I select TNS from the connection type dropdown list , I get a message below that states: 'Warning ORACLE_HOME environment variable is not set'.
    This is odd, since I use SQL Developer daily on this same system with no issues. I have checked with regedit and the ORACLE_HOME environment variable is indeed set.
    Any ideas?

    I ran into a similar where Data Modeler was finding the TNSNAMES.ORA file for the older Oracle 8i client (which I have to keep installed because some older legacy software depends on it) rather than the 10g client. My solution was to make use of a suggestion for a simialr problem with SQL Developer and create a DOS "cmd" file in the "+<data modeler install dir>+\datamodeler" directory that sets the ORACLE_HOME environment variable to my 10g client directory and then runs "datamodeler.exe". My file is named "datamodeler.cmd" and contains the following two lines:
    SET ORACLE_HOME=C:\oracle\product\10.2.0\client_1
    start datamodeler.exe
    HTH.
    Ed. H.

  • Help needed to mark or highlight certain columns in Data Modeler

    Hi there,
    We use the data modeler to present our data model to the security department to explain/prove for which columns we ran a data masking/anonymization function. I am looking for any kind of way to either highlight or change a set of columns across all tables. So far I could only identify the red dot for “not null” and the option to introduce a ADT/UDT like “ANONYMIZED” to show in the graphical representation what we have done. As we base our DM on empty tables for decent overviews including the PK/FK relationships we are quite flexible. Still the “misusage” of ”Not Null” or ADT/UDT columns does not go together with the required PK/FK’s. Does anyone know a way to mark or highlight certain columns automatically when importing a data dictionary? (Unfortunately RDBMS comments do not get included in any overview). Thanks a lot. Cheers Stefan

    Hi Philip, Thanks a lot for puting this enhancement request through.
    Just downloaded the latest Patch upgrading to v 3.1.2-704 and confirmed that this functionality is not available yet.
    Keeping your experience in mind what kind of expectation to you have for the approval and realization of your request?
    Most likely this will take a couple of month – right? Or is there a beta version of 3.2 already available we could use.
    Thanks a lot. Cheers Stefan

Maybe you are looking for