Advice on a data model

I am using SP Server 2010.  The requirements of a project are to:-
Search an external database so that we can bring back rows from that database with say 5 fields.
Create a list in SP to hold about 25 site columns
Use the list in (2) to add rows sourced from (1) and fill in the site columns specified in (2)
Poll the (1) to check if new rows have been added and, if they have, add them to the list in (2)
Could anyone please advise upon the best way to model these requirements in SP given this MSDN
documentation?
Using BCS presumably?
Should we favour External Columns in (2) as opposed to lookups for example (see my
post here)?
We are basically trying to convert an Excel Workbook to SP. If we have the option (and we may not) is Excel Services the best way to do this?
What is List Pattern we should adopt?
Thanks,
Maz

Hi,
For your first requirement, yes, you can use business connectivity service to bring rows from the external database field.
You can refer the detailed article below:
Using Business Connectivity Services in SharePoint 2010
For your second and third requirement, yes you need to create external data column to fill the external database field data to the list in the list. As external list has limitation as the source of look up column directly.
More information about external data column:
Create a SharePoint List with External Data Columns
If your requirement "convert excel workbook to SP", yes, you can use excel service to display workbook content in browser directly.
More information:
Excel Services in SharePoint Server 2010
For your fourth requirement, you can create a console application with Server Object Model to check the data in the external list in first requirement, if there is data in the list ,then you can add the data to external columns in requirement 2.
Programmatically set external column
For the list pattern, it will depend on your business logic and you data count,if your data is very large, then you can use the large list pattern.
Thanks
Best Regards
TechNet Community Support
Please remember to mark the replies as answers if they help, and unmark the answers if they provide no help. If you have feedback for TechNet Support, contact
[email protected]

Similar Messages

  • How to print Data Model in Report??

    Do anyone know how to print Data Model??
    Thanks your help.
    - Frank

    Hi Venkat,
    Thanks alot.... i am checking the appendix k.
    I want to know that if i want to customize the report as per the client requirement,
    can you please let me know which template i will use if i required remittence advice as well as check print data on the layout.
    The following is the list of field i required:
    Vendor ID
    Check Date
    Check number
    invoice date
    invoice/ CR memo number
    invoice description
    invoice gross amount
    invoice discount amount
    invoice net amount
    total gross amount
    total discount amount
    total net amount
    logo
    company name and info
    bank name and info
    check number
    check amount spelled out
    check date
    check amount numeric
    payee name
    CEO signature
    MICR check number
    MICR routing number
    MICR bank account number
    mailing return address
    mailing address
    Venkat it would be great help if you please let me know the exact process to be follow for check printing report.
    I worked on bi publisher report in 11i, but in 12i the process is bit different.
    also if you have any template ready can you please send me on my mail id ... i forwarded you the test mail from my official id or on [email protected]
    Looking forward to your kind response.
    Regards
    Ratnesh

  • Oracle BI Publisher 11g - Known Data Model Limitations

    Hi
    I'm new to BI Publisher 11g. I'm attempting to find out what the known limitations of BI Publisher 11g are.
    In particular, I would like to know the maximum number of queries allowed per data model or if a report may be based on multiple data models.
    Scenario:_
    My organization requires an electronic reporting book (PDF document) effectively consisting of *40-50* data tables and graphs. Someone has sold the idea to IT and the business that BI Publisher as the tool that deliver this type of functionality. Having used the tool now for the last couple of days, it's starting to look it's not the case.
    How do i achieve my ultimate goal of combining disparate report output into 1 huge document?
    Do I develop each reporting section as a separate data model/report pairs outputting XML files and then combine all XML files into 1 target document?
    If anyone can please give me some advice on how to go about this, it would be greatly appreciated.
    Thanks
    Greg

    This is related to browser, try it in some other browser.or else check ur bi publisher settings and data source configurations. Mostly his wil happen if u changed some thing in blaf rich plus.css changes

  • SQL Developer Data Modeler with TFS?

    I'm looking for a way to hook up Oracle SQL Developer Data Modeler to Team Foundation Server for version/source control. We currently run TFS for all source control repositories and want to avoid installing subversion just for the Data Modeler product.
    Any suggestions would be great

    Unfortunately TFS isn't currently supported natively in the tool, although maybe someone else who has manually set it up for manage models will have some advice for you.

  • Inheritance in OPA data model

    Hi all,
    I have a problem which OPM does not seem to handle well, but maybe I am lacking knowledge concerning some functionality. I am using a client patched 10.1 version of OPM.
    Situation:
    - I have an attribute in Global, which states the possibility to export an entitlement. This attribute is deducted using multiple compositions of conditions on both "child" level and "applicant" level. These are both entities in my data model.
    - However, some of the conditions are the generic for both roles. Therefore, what I really want to do is check attributes on "person" level and relate these to "the child" and to "the applicant".
    - Now, I've created an entity "the person" with a relation to "the child" and a relation to "the applicant".
    This raises an issue with the For/ForAll operator
    Issue:
    - To use the "For" operator, a -to-one relationship is needed. However, I can have two persons of which one is a child and one is an applicant. So, I need a - to-many relationship for person (from Global)
    - If I use the "ForAll" operator to 'work around' this problem, the goal attribute will stay inconclusive as the relationship from person 1 to the child is known, but from person 2 to the child in unknown (and vice versa for the applicant)
    Question:
    It seems to me that this is an issue of inheritance which OPM cannot handle. I am wondering:
    - has anyone dealt with this problem and/or does anyone have a solution?
    - is it likely that the entity containment functionality of OPM 10.2 + will be able to solve my issue?
    Thanks & regards, Els

    Given that I don’t know the details of your source material, and I don’t know Dutch, perhaps the best option is if I show you how to get a simple inferred relationship set up and working. Hopefully then you’ll see what it’s doing and know how to set it up for your own rulebase.
    1. Create a new empty test rulebase.
    2. Add a Properties file. Add a regular one-to-many containment relationship from Global to ‘the person’.
    3. From within the Properties file, have Global selected and in the right hand pane select the Relationships tab.
    4. Right-click in an empty space in the right hand pane and select ‘New Relationship’. Use these settings:
    - Source: Global (This will already be set, and will be read-only. If it isn’t already set to Global, then you didn’t do Step 3.)
    - Target: the person
    - Relationship type: Inferred (many-to-many)
    - Text: the adults
    5. Save the Properties file.
    6. Create a Word rule doc.
    7. Add a membership rule for the inferred relationship. A membership rule specifies the logic used to decide which instances are members of the inferred relationship. The parts in italics below are just examples of what the red configuration text will look like when the rule is compiled.
    +[IsMemberOf(“person”,adults)]+ the person is a member of the adults if
    +[p1 >= 18]+ the person's age >= 18
    In my example, I’m just basing it on age. It could be based on just about anything. Try to name your inferred relationship appropriately. I called it “the adults” since that’s what the group is. If I needed to have a group of people >= 18 years and who are female, then I’d call it something like ‘the female adults’, and obviously I’d include gender in the membership rule.
    8. Add some simple entity calculation rule such as these:
    the number of adults in the group = InstanceCount(the adults)
    the total amount of savings of the adults in the group = InstanceSum(the adults, the person's savings amount)
    9. Build and Debug. Work in the Data tab, not the OWD tab. You’ll be able to see more detail about what’s going on if in the Data tab.
    10. Add three instances of ‘the person’. Make two people older than 18 years, and one person younger than 18 years. Give each of them a different savings amount.
    11. Look at the Global results and you will see that the InstanceCount rule calculated ‘2’ and the InstanceSum rule only added up the savings from the subset of people who are adults.
    My general advice on this is:
    If the logic you need to do over subsets of entity instances is relatively simple, and can be easily accommodated by existing Instance functions, then don't bother with inferred relationships. Just set up the regular containment relationship and use the existing entity functions. Here are few examples of less complex entity calculations which can easily be handled without inferred relationships:
    - count the number of adults
    - calculate the total income of all household members who are over 25 years
    - find the age of the youngest male child
    So my example rulebase above is actually one for which I wouldn’t use inferred relationships if that was all I needed it for. I just wanted to pick a simple example to explain the concept. To do the example above without inferred relationships, you’d just need to use InstanceCountIf and InstanceSumIf and put the conditional logic in each calculation rule.
    If the logic you need to do over subsets of entity instances is much more complex and either it's not possible without inferred relationships, or it's just easier/more concise with inferred relationships, then add inferred relationships.
    Cheers,
    Jasmine

  • Data Model best Practices for Large Data Models

    We are currently rolling out Hyperion IR 11.1.x and are trying to establish best practces for BQYs and how to display these models to our end users.
    So far, we have created an OCE file that limits the selectable tables to only those that are within the model.
    Then, we created a BQY that brings in the tables to a data model, created metatopics for the main tables and integrated the descriptions via lookups in the meta topics.
    This seems to be ok, however, anytime I try to add items to a query, as soon as i add columns from different tables, the app freezes up, hogs a bunch of memory and then closes itself.
    Obviously, this isnt' acceptable to be given to our end users like this, so i'm asking for suggestions.
    Are there settings I can change to get around this memory sucking issue? Do I need to use a smaller model?
    and in general, how are you all deploying this tool to your users? Our users are accustomed to a pre built data model so they can just click and add the fields they want and hit submit. How do I get close to that ideal with this tool?
    thanks for any help/advice.

    I answered my own question. in the case of the large data model, the tool by default was attempting to calculate every possible join path to get from Table A to Table B (even though there is a direct join between them).
    in the data model options, I changed the join setting to use the join path with the least number of topics. This skipped the extraneous steps and allowed me to proceed as normal.
    hope this helps anyone else who may bump into this issue.

  • Need help with Data Model for Private Messaging

    Sad to say, but it looks like I just really screwed up the design of my Private Messaging (PM) module...  *sigh*
    What looked good on paper doesn't seem to be practical in application.
    I am hoping some of you Oracle gurus can help me come up with a better design!!
    Here is my current design...
    member -||-----0<- private_msg_recipient ->0------||- private_msg
    MEMBER table
    - id
    - email
    - username
    - first_name
    PRIVATE_MSG_RECIPIENT table
    - id
    - member_id_to
    - message_id
    - flag
    - created_on
    - updated_on
    - read_on
    - deleted_on
    - purged_on
    PRIVATE_MSG table
    - id
    - member_id_from
    - subject
    - body
    - flag
    - sent_on
    - updated_on
    - sender_deleted_on
    - sender_purged_on
    ***Short explanation of how the application currently works...
    - Sender creates a PM and sends it to a Recipient.
    - The PM appears in the Sender's "Sent" folder in my website
    - The PM also appears in the Recipient's "Incoming" folder.
    - If the Recipient deletes the PM, I set "deleted_on" and my code moves the PM from Recipient's "Inbox" to the "Trash" folder.  (Record doesn't actually move!)
    - If the Recipient "permanently deletes" the PM from his/her "Trash", I set "purged_on" and my code removes the PM from the Recipient's Message Center.  (Record still in database!)
    - If the Sender deletes the PM, I set "sender_deleted_on" and my code moves the PM from the Sender's "Sent" folder to the "Trash" folder.  (Record doesn't actually move!)
    - If the Recipient "permanently deletes" the PM from his/her "Trash", I set "sender_purged_on" and my code removes the PM from the Sender's Message Center.  (Record still in database!)
    Here are my problems...
    1.) I can't store PM's forever.
    2.) Because of my design, the Sender really owns the PM, and if I add code to REMOVE the PM from the database once it has a "sender_purged_on" value, then that would in essence remove the PM from the Recipient's Inbox as well!!
    In order to remove a PM from the database, I would have to make sure that *both* the Recipient has "purged_on" value and the Sender has a "sender_purged_on" value.  (Lot's of Application Logic for something which should be simple?!)
    I am wondering if I need to change my Data Model to something that allows my autonomy when it comes to the Sender and/or the Recipient deleting the PM for good...
    One the other hand, I believe I did a good job or normalizing the data.  And my current Data Model is the most efficient when it comes to saving storage space and not having dups.
    Maybe I do indeed just need need to write application logic - or a cron job - which checks to make sure that *both* the Sender an Recipient have deleted the PM before it actually flushes it out of my database to free up space?!
    Of course, if one party sits on their PM's forever, then I can never clear things out of my database to free up space...
    What should I do??
    Some expert advice would be welcome!!
    Sincerely,
    Debbie

    rp0428,
    I think I am starting to see my evil ways and where I went wrong... 
    > Unfortunately his design is just as denormalized as yours
    I see that now.  My bad!!
    > the last two columns have NOTHING to do with the message itself so do NOT belong in a normalized table.
    > And his design:
    >
    > Same comment - those last two columns also have NOTHING to do with the message itself.
    Right.
    > The message table should just have columns directly related to the message. It is a list of unique messages: no more, no less.
    Right.
    > Mark gave you hints to the proper normalized design using an INTERSECT table.
    > that table might list: sender, recipient, sender_delete_flag, recipient_delete_flag.
    > As mark suggested you could also have one or two DATEs related to when the delete flags were set. I would just make the columns DATE fields.
    >
    > Once both date columns have a value you can delete the message (or delete all messages older than 30+ days).
    >
    > When both flags are set you can delete the message itself that references the sender and the message sent.
    Okay, how does this revised design look...
    MEMBER --||-----0<-- PM_DISTRIBUTION -->0-------||-- PRIVATE_MSG
    MEMBER table
    - id
    - email
    - username
    - first_name
    and so on...
    PM_DISTRIBUTION table (Maybe you can think of a better name??)
    - id
    - private_msg_id
    - sender_id
    - recipient_id
    - sender_flag
    - sender_deleted_on
    - sender_purged_on
    - recipient_flag
    - recipient_read_on
    - recipient_deleted_on
    - recipient_purged_on
    PRIVATE_MSG
    - id
    - subject
    - body
    - sent_on
    Is that what you were describing to me?
    Quickly reflecting on this new design...
    1.) It should now be in 3rd Normal Form, right?
    2.) It should allow the Sender and Recipient to freely and independently "delete" or "purge" a PM with no impact on the other party, right?
    Here are a few Potential Issues that I see, though...
    a.) What is to stop there from being TWO SENDERS of a PM?
    In retrospect, that is why I originally stuck "member_id_from" in the PRIVATE_MSG table!!  The logic being, that a PM only ever has *one* Sender.
    I guess I would have to add either Application Logic, or Database Logic, or both to ensure that a given PM never has more than one Sender, right?
    b.) If the design above is what you were hinting at, and if it is thus "correct", then is there any conflict with my Business Rule: "Any given User shall only be allowed 100 Messages between his/her Incoming, Sent and Trash folders."
    Because the Sender is no longer "tightly bound" to the PRIVATE_MSG, in my scenario above...
    Debbie could send 100 PM's, hit her quota, then turn around and delete and purge all 100 Sent PM's and that should in no way impact the 100 PM's sitting in other Users' Inboxes, right??
    I think this works like I want...
    Sincerely,
    Debbie

  • Excel Workbook with Excel Data Model stored on SharePoint - daily refresh

    I'm not sure if this is exactly the right forum but I'm hoping that someone here can either answer or point me in the right direction.
    I have an Excel Workbook with an Excel Data Model.
    The Excel Data Model uses SQL to contact our data warehouse and pull through data into tables.
    We are storing the Workbook on a SharePoint site and viewing it using Excel Services.
    The data in our data warehouse is updated daily and so I would like to refresh the workbook daily and remove the option to refresh the data in the browser.
    Is this possible with a workbook that has an Excel Data Model (I've seen lots of posts that relate to workbooks with connections to tabular models).
    Thanks
    Paul

    Hi Paul,
    I have answered this issue in this thread that you posted in SQL Server PowerPivot for SharePoint forum, please see:
    http://social.msdn.microsoft.com/Forums/en-US/9627939c-e9f1-48ae-a6ed-0c238d8f2d69/excel-workbook-with-excel-data-model-stored-on-sharepoint-daily-refresh?forum=sqlkjpowerpointforsharepoint
    Furthermore, this issue is more related to PowerPivot for SharePoint. If you have any more questions about PowerPivot for SharePoint, I would sugget you open a new thread in the forum below for better support:
    http://social.msdn.microsoft.com/Forums/en-US/home?forum=sqlkjpowerpointforsharepoint
    Thanks for your understanding.
    Regards,
    Elvis Long
    TechNet Community Support

  • Issue while creating Data Model in BI Publisher and logging into xmlpserver

    Hi All,
    We are facing an issue in OBIEE 11.1.1.5.
    If we are logging with Non Admin Id (other than weblogic) and select New Data Model, a blank screen is coming. Where as, if we use Admin Id, we are getting screen as usual for creating data model. There were some blogs mentioning to change Priviledges for  SOAP access, but that approach is also not working.
    Further, we are also not able to open xmlpserver with Non Admin Id.
    Any help or pointers would be great.
    Regards,

    how about pasting the content of your data template here, so that forum members can see what could be the problem.

  • Data Modeler : Modifying the Table Report layout

    Hi ,
    I'm using SQL Data Modeler (DM) 4.0 EA .
    I used the File -> reports option to generate the reports on Table and in the report
    1) I couldn't see the column data type ? (Refer below table ) , Is there any options to bring on the Data Type in the Report ?
    No
    Column Name
    PK
    FK
    M
    Data Type
    DT
    kind
    Domain Name
    Formula
    (Default Value)
    Security
    Abbreviation
    1
    ID1
    P
    Y
    (10)
    LT
      2
    ID2
    F
    Y
    (10)
    LT
    3
    ID3
    Y
    (1)
    LT
    4
    ID4
    Y
    (10)
    LT
    5
    ID5
    F
    (20,4)
    LT
    2) In options to manage the Table report (Reports -> Manage). Currently I can only remove a heading (like Descriptions Notes ,columns,column comments.). Is there is any way to customize the report in such a way that I can remove some columns (mostly the empty ones) from the report and add the column comments along with the column table than a new column comments table.
    Please let me know  ,is there is an option to customize the report .
    Note: I even tried generating the report using the search then report , but it doesn't give a complete report (i was only able to generate a report on column or table name or constraints , not all together.)
    Thanks,
    Srinivasan.K

    Thanks for the immediate reply.
    I checked on the DM again, the report comes good for the one which has the data type . But now all my data type changed to UNKNOWN and lokks like ,due to this my table report is coming without the "Data Type".
    Can you please help me in fixing this unknown data type issue ?
    Table report for the ones with proper Data type:
    Columns
    No
    Column Name
    PK
    FK
    M
    Data Type
    DT
    kind
    Domain Name
    Formula
    (Default Value)
    Security
    Abbreviation
    1
    ID1
    Y
    NUMERIC (10)
    LT
    2
    SCRIPT
    Y
    VARCHAR (1024)
    LT
    3
    UPGRADE_S
    Y
    VARCHAR (10)
    LT
    4
    UPGRADE_D
    Y
    VARCHAR (250)
    LT
    5
    UPGRADING_F
    Y
    VARCHAR (20)
    LT
    6
    TIMESTAMP2
    Y
    Timestamp
    LT

  • Can "trigger" management be addressed in the data modeler?

    Please add an easier way to add/modify triggers.
    I reverse-engineered my model into the data modeler and my triggers all migrated correctly to the model but, for the life of me, had no idea how to view, modify or create triggers.  After extensive research, I found that they can be accessed in the left-hand Browser by drilling some 8 levels to view the triggers for each table - why?  Why can't this functionality be added to properties context menu for the table.  I'm not a big advocate for triggers but if I need to work on one, the current methodology is tedious at best.
    Of course, if there's some other way to do this, please inform...
    Accessing Triggers via the Browser
    Design -> Relational Model -> APEX_DEMO -> Physical Model -> Oracle Database 11g -> Tables -> ACTIVITY_LOG -> Triggers

    Thanks Joop and Phillip...
    While I occasionally use SQL Developer, I currently use PL/SQL Developer (Allround Automations) as my development tool of choice and bringing a 3rd tool as SQL Developer into the mix, makes no sense to me.
    I completely agree that managing a trigger via a development tool is much easier but one occasionally wants to quickly add a trigger either manually or via a template and this is not easy currently.
    While functions, procedures, package, etc. are all important to a data model, the existence of most triggers are completely dependent on a table and are a critical component in any model that uses them.  The ability to access and modify them easily in Data Modeler is important and shouldn't be an afterthought as is perceived here.  If trigger access cannot be added to the Table Properties (as noted in the image below), at the very least, please provide a shortcut methodology to do so.

  • Not able to save a newly created Data Model in BI Publisher.

    Hello everyone,
    I was trying to create a new Report (my very first report) in BI Publisher. I created a new Data Model and a Data Set. But I am not able to save the Data Model and thus not able to proceed to the report. There's no error shown. Same happens with reports too.
    Kindly provide me with a solution.
    Regards
    Ash
    Edited by: 902739 on 28-Dec-2011 01:02

    You should have a group called 'XMLP_ADMIN' in your repository and administrator or the user you are using should be part of that group. Look into paint.rpd for all these.

  • Table or view does not exist - Data Dictionary Import Wizard(Data Modeler)

    Hi All,
    In Data Modeler, Data Dictionary Import Wizard, I'm able to connect to database.But while going to the second stage (Select Schema/Database), I'm getting an error "ORA-00942: table or view does not exist".
    I am able to select the table with select * from all_tables and I can open many tables as well.
    Could anyone tell me, whether I'm missing any privilege, that causing this error.
    Thanks in advance for you support.
    Thanks.

    Hi,
    Thanks for your response, sorry for my late reply as I was away from my place.
    Yes, it is showing "Connection established successfully".
    log file as below-
    2012-08-02 10:37:26,471 [main] INFO ApplicationView - Oracle SQL Developer Data Modeler 3.1.1.703
    2012-08-02 10:39:42,889 [AWT-EventQueue-0] ERROR AbstractDBMExtractionWizardImpl - java.sql.SQLSyntaxErrorException: ORA-00942: table or view does not exist
    Pls see oracle version details-
    Oracle Database 11g Enterprise Edition
    11.1.0.6.0
    64bit Production
    Thanks again and waiting for reply.

  • Copy data models and reports from BW 3.1 to NW 2004s

    Hi experts,
    Our client has two BW servers: BW 3.1 and BI 7. BW 3.1 contains lots of data models and reports. And the BI 7 server is newly installed.
    Now we want to copy these data models and reports from BW 3.1 to the new BI 7 server. Are there any solutions for this?
    Thank you very much in advance.

    Hi Frank,
    Sounds like a cross version transport is needed.
    This is a solution we have used to do what you want to do:-
    Create and release a transport as per normal.
    Copy and transport the files from the source system (BW 3.1) e.g /usr/sap/trans/data & /usr/sap/trans/cofiles to the same folders on the target system.
    Basis help is needed here.
    From here onwards using stms_import should help you in the normal manner.
    Works a treat.
    Have transported the following all correctly appearing as 3.x data models in NW2004s.
    DSO objects.
    Cubes
    Transfer/Update rules
    Reports.
    Cheers,
    Pom

  • Some observations on my first use of Data Modeler Beta

    First of all, I can see this tool has a lot of promise.
    I hope Oracle keeps at it, it could turn into a real winner if all the features I see being worked on mature.
    Thanks!
    Here are a few observations on things that I found non-obvious or tedious to do.
    1. When designing an entity, I want to give it a name, a definition, attributes and keys. I want that process to be quick and require the minimum amount of mouse-clicking/navigation fiddling as possible. The current way of defining the attribute's datatype and size is painfully slow. I have to click to get a pop-up. Then I have to click to choose from a set of categories. Then I have to click on a dropdown list. If I try to use the down-arrow on the dropdown list, it works, but not if I go past the one I want. The up-arrow won't take me backwards in the list, so more clicking. It's just a nasty, slow interface to do a simple task that I have to do a thousand times in a data model. If I need to change the size of something, back I go thru the entire process all over again.
    That makes it doubly slow to work in the most natural way, which is to list the attributes and the datatypes, then come back and refine the sizes once the model is maturing and relatively stable.
    2. Adding an additional attribute requires a mouse click instead of a down-arrow. That means I have to take my hands off the keyboard to add a new attribute. Maybe there is some short-cut key that does that, but I have better things to do than memorize non-standard keyboard mappings. Make the down/up arrows works as they should.
    3. Adding the comment that describes the attributes is not quite as slow, but still requires more keystrokes/mouse movements than it should. It's hard enough to get developers to document their attributes, don't discourage them.
    4. I can't see the list of attributes, data types, sizes, key/mandatory settings, and the comment at one time while editing the entity. Makes it harder to grasp what all the attributes mean at a glance, which slows down the modeling and the comprehension of an existing model.
    5. All the entities I created had primary keys with columns in them. But when I tried to have it build a physical model, it complained that some relationships had no columns in them. For the life of me, I couldn't figure out how to fix that. Never had that problem in any other case tool.
    6. Getting it to generate DDL was awkward to find. Make it something obvious, like a button on the toolbar that says "Generate DDL".
    7. Apostrophes in the Comments in RDBMS are not escaped, so the generated DDL won't run.
    8. For the ease of use/speed of use testing on high-volume key tasks, make the developers do the task 1000 times in a row. Make them use long names that require typing, not table A with columns c1, c2 and c3. Long before they get to iteration 1000 they will have many ideas on how to make that task easier and faster to do.
    9. Make developers use names of things that are the maximum length allowed. For example, for a table name in oracle, the max length of the name is 30 characters. The name of one testing table should be AMMMMMMMMMMMMMMMMMMMMMMMMMMMMZ. That's a capital A followed by 28 capital M's and a capital Z. For numbers, use the pattern 1555559. If the developers can't see the A and Z or 1 and 9 in the display area for the name in the default layout for the window, they did the display layout wrong. For places where the text can be really long, choose a "supported visible length" for that field and enter data in the pattern AMMMMMMMMMMMMMMMQMMMMMMMMZ, where Q is placed at the supported visible length. if the A and Q don't show, the layout is wrong.
    10. SQL Developer has quite a few truly gooberish UI interaction designs and I can see some of that carrying over to the Data Modeler tool. I really recommend getting a windows UI expert to design the ui interface, not a java expert. I've seen a lot of very productive windows user interfaces and extremely few java interfaces suitable for high-speed data entry. Give the UI expert the authority to tell the java programmers "I don't want to hear about java coding internals - make the user interface perform this way." I think the technical limitations in java UIs are much less than the mindset limitations I've seen in all to many programmers. That, and making the developers use their code 1000 times in a row to perform key tasks will cause the UI to get streamlined considerably.
    Thanks, and keep up the good work!

    Dear David,
    Again thank you for your valuable and highly appreciated feedback. Find included a more elaborated answer to your observations:
    *1. When designing an entity, I want to give it a name, a definition, attributes and keys.*
    In the new Early Adopter Release, a "SQL developer like" property window has been added for Entities and Tables. For each attribute/column you will have the ability to add name, datatype, Primary Key, Mandatory and comment from one and the same screen
    *2. Adding an additional attribute requires a mouse click instead of a down-arrow.*
    An enhancement request has been created
    *3. Adding the comment that describes the attributes is not quite as slow, but still requires more keystrokes/mouse movements than it should.*
    In the new Early Adopter Release, a "SQL developer like" property window has been added for Entities and Tables. For each attribute/column you will have the ability to add name, datatype, PK, M and comment from one and the same screen
    *4. I can't see the list of attributes, data types, sizes, key/mandatory settings, and the comment at one time while editing the entity. Makes it harder to grasp what all the attributes mean at a glance, which slows down the modeling and the comprehension of an existing model.*
    See former answers. For meaning of attributes you can also use the Glossary and Naming Standardization facilities: see Tools Option menu, Glossary and General Options for naming standards
    *5. All the entities I created had primary keys with columns in them. But when I tried to have it build a physical model, it complained that some relationships had no columns in them. For the life of me, I couldn't figure out how to fix that. Never had that problem in any other case tool.*
    A Bug report has been created. Issue will most probably be solved in the nexr Early Adopter release
    *6. Getting it to generate DDL was awkward to find. Make it something obvious, like a button on the toolbar that says "Generate DDL".*
    An enhancement request has been created.
    *7. Apostrophes in the Comments in RDBMS are not escaped, so the generated DDL won't run.*
    A bug report has been created
    *8. For the ease of use/speed of use testing on high-volume key tasks, make the developers do the task 1000 times in a row. Make them use long names that require typing, not table A with columns c1, c2 and c3. Long before they get to iteration 1000 they will have many ideas on how to make that task easier and faster to do.*
    I aplogize, but I don't understand clearly what you want to say with the use/speed of use here.
    *9. Make developers use names of things that are the maximum length allowed.*
    Our relational model is for use for not just Oracle, but also DB2, SQL Server and in the future maybe other database systems. Whicjh means that we can't taylor it to just one of these database systems. However you can set maxinum name lenghts by clicking right on the diagram and select Model Properties and here you can set naming Options. Here you can also use the Glossary and Naming Standardization facilities: see Tools Option menu, Glossary and General Options for naming standards
    *10. SQL Developer has quite a few truly gooberish UI interaction designs and I can see some of that carrying over to the Data Modeler tool.*
    Fully agree. As you will see in our next Early Adopter release we have started to use SQL Developer like UI objects.
    Edited by: René De Vleeschauwer on 17-nov-2008 1:58

Maybe you are looking for

  • Why isn't the overlay working?

    hey everyone - i've got a clip in final cut pro that i wanted to end with a quick light streak overlay. i opened up motion, made a 5 second black 0% opacity sequence, and added the streak. then clicked the layer, and chose the blending mode of "overl

  • Stopping RAC instances on shutdown (10g - RHEL3)

    Any advise on the correct way to script instance stop in cluster nodes? As soon as a node comes up, CRS automatically starts its instances. However, according to alert logs, they are not being cleanly stopped (if left as configured by the installatio

  • Nullpointer when creating JFileChooser

    I am launching a regular java app in a Desktop I created ( CreateDesktop() ) on a Windows 7 64 bit box. The code is all compiled in 32 bit, and the JVM being used to launch the app is also 32 bit. IN this application I create a JFileChooser object, J

  • MAC mail bouncing

    Ever since the MAC mail went down 3 weeks ago, I have experienced problems: A lot of incoming mail is bouncing. And tonight, one of my accounts is having trouble sending mail. I keep getting the pop up window that says the server timed out. After 4 o

  • Converrsion of .class to exe file

    i googled the above problem. in one of the forms i got the information. it instructed me to use javaw [filename] to acheive the same. but when i use javaw [ file name] it comes with the msg "cannot find main class. program will exit". I just want to