High level design options

Hi, I'm trying to decide on the best way to desing an application that will be responsible for the following:
-Display and control of 5 temperature zones (analogue/digital)
-Display and control of 4 Mass Flow Controllers (analogue)
-Control of various serial devices
-Logging and other standard application features
I have no problem writing small VI's do control one of the temp zones or one of the MFC's but putting it all together is proving to be harder than I expected. I am using LV 7.1 with PCI 6024E and PCI 6602 DAQ cards. I've looked at application examples like the 'Cookie factory' which have been very useful but I'm worried about how to do all the I/O.
I've noticed that if I write a separate VI to control each temp zone, they don't seem to all be able to access the hardware at the same time (not surprising) but does this mean that I'll have to do all my I/O at the same time and therefore in the same VI?
I've also thought about using the DAQopc server in conjunction with the DSC module which would allow my VI's to just read and write to tags as apposed to trying to read directly from the DAQ. Does the DAQopc server work with DAQmx? I haven't been able to see any of my DAQmx tags in the opc server.
Any help would be much appreciated. Thanks.

I do not have any experience with the DSC stuff, so I will not comment on that. I had delayed responding, thinking that someone with more experience along those lines might make some suggestions. Anyway with the cautions out of the way, here goes my opinions.
I like state machine architectures. In particular I have separate, parallel state machines for DAQ, GUI, and data processing. I use queues and functional globals to transfer data and commands among the state machines. In your case I might further subdivide the DAQ into a machine for each of the cards and another for the serial communications.
For example if all the temperature zones (A, B, C, D, E) were monitored and controlled through the PCI 6024, I might have a loop which reads and writes to the device (PCI 6024). Inputs would be commands sent over a queue and would be of the form "Read Temperature " or "Set Temperature <150>" where the part in <> brackets is a parameter. The output (in separat...

Similar Messages

  • High-level design for business process scenario

    Hi all, I have a business process scenario for which CRM and ECC has to be used. The scenario is like this:
    There are 2 subsidiaries of a company - A and B, which are into trading of materials.
    Between A and customer there is sales agreement. Between B and vendor there is purchase agreement.
    Whenever customer requires materials, it will send request to A. So, A creates a sales order. B in turn sends a request to vendor for which a purchase order is created. The vendor ships the goods directly to customer. The vendor bills B for the purchase order. B adds some margin and then it goes for inter-company settlement with A. A bills the customer for the materials. In this entire process, both A and B earns profit through margins.
    It has to be noted that CRM has to be used as contact point with customer.
    Now I have the following questions:
    1. How would the high-level business design be like?
    2. How would be the document flow?
    3. How can inter-company settlement and billing be done?
    All constructive comments and feedback will be rewarded.

    Hi Animesh,
    The thumb rule to CRM process designing is stick to customer process...
    Thus u jst hav to stick to process design of company A and its interaction with Customer.
    We are not concerned with how the material will be purchansed,but we are concerned with how we will be delivering it to end customer.
    Thus, once customer place an order to company A, our process would suggest a sales order for the customer with one of the partner function "Vendor" as B with the required material. Now later process in R/3 would convert this sales order with customer as soldtoparty to purchase order with vendor as B and one of the secondary partner soldto party as our original customer.
    And billing would happen the way it is... ie. based on sales order.. thus u will have some inflow.. again u will be raising a purchase requisition to company B thus that would be your outflow..
    That was a really good scenario to think about.
    Best Regards,
    Pratik Patel
    <b>
    Reward with Points!</b>

  • Needed help in high level design approach of this service

    I have a two Proxy services proxy service A and Proxy Service B
    Proxyservice A takes the XML request from the consumer and puts the message in JMS queue(queueA) and gives an immediate acknowldgment. Proxyservice B(Async) picks that message from queue and invokes the provides, if we get any binding fault or remote fault 5 retries are made (Proxy service XA enabled and if proxy service gets errored out then message remains in same queue) and message is put into error queue( if we get any business fault the message is put into undelivered queue no retries needs to be made). So whenever the message is put into the errorqueue an email notification must be sent to the sustain team. And then the ,message needs to be moved back to the queueAso that the message gets processed again
    Your help is appreciated. Need urgent help on this.
    Thanks,

    I have a two Proxy services proxy service A and Proxy Service B
    Proxyservice A takes the XML request from the consumer and puts the message in JMS queue(queueA) and gives an immediate acknowldgment. Proxyservice B(Async) picks that message from queue and invokes the provides, if we get any binding fault or remote fault 5 retries are made (Proxy service XA enabled and if proxy service gets errored out then message remains in same queue) and message is put into error queue( if we get any business fault the message is put into undelivered queue no retries needs to be made). So whenever the message is put into the errorqueue an email notification must be sent to the sustain team. And then the ,message needs to be moved back to the queueAso that the message gets processed again
    Your help is appreciated. Need urgent help on this.
    Thanks,

  • High Level Network Design

    Hi Guys
    I am posting this because I am starting my career into network design and want some help in it. I am at present in need of a high level design overview as I need to prepare some high level network design documents. Can anyone shower some thoughts in it as how about doing this and if any there is a template for HDD so that it maybe useful.
    Also I believe in keeping information as transparent as possible to the readers of the document and need someone to explain in very simple terms if at all it is possible.
    Thanks a lot
    Vin

    Hi Vin,
    I would check the Cisco SBA and Validated Design Zone as a first pass.
    Lots of great design documents there.
    As for how I would create a high level design - keep it simple.  You just want an overview of the connectivity - e.g. for a dual-site head office with 100+ branch wan, I would only show a single branch site as a template.
    Every network is different, but the more documentation you write and read the more you will define your own style.
    Apologies I can't give you any of my customer's documentation - NDA's and everything!
    Regards, Ash,

  • How to pass High Level Indicator (HLI) for Components?

    Purchase Order has to pass from one system 1 (SAP) to another system 2 (SAP) and to create Sales Order in the  system 2 automatically through ALE IDOCs.
    How to pass HLI from one system to another using IDOCS?  Which segments need be passed and how?  help me with this. Thank you.

    Hello Cnu,
    Welcome to SDN.
    High Level Design gives you an overview of what you are going to do to get the desired result (without going deep technical config. details)
    Eg: What interfaces are you going to use ? Where you are going to load the data(including staging) ? How the reports are delivered ? How the Application Support is provided ? etc..
    For a generic description of High Level Design see this http://en.wikipedia.org/wiki/High-level_design
    by the way, who is your client in Aus. I am from Syd.
    Cheers
    Praveen

  • Design considerations for wrapping higher level process

    So Ive used workflow now on a few projects and have become comfortable building/coding with it.  Recently a discussion was started to design a new process to automate a part of the business Im in and there are several different departments, each with
    their own different functions and processes that they perform.  During this discussion, the thought came to me - why stop with just the "small" picture where I was using WF to contain and execute business rules, but also wrap the higher level
    abstractions also.
    For example, a department performs steps A, B and C.  Each step might have some branching in it (if/then etc).  Its only the "steps Im using workflow currently.
    Above the department, there is another layer, lets call it "group".  There would be several different groups, each with its own departments, some with the same name as other groups, but the internal structure might be different for a given
    department between groups.  So this means that if I modeled a department in one group, I cant assume I can reuse that department in another groups workflow.
    My main unknown however, is how would a user interact with a workflow that was this high level.  Abstracting higher levels, means users dont need to interact with or execute the whole workflow each time, only a part of it.  So am I looking at pausing/resuming
    the bigger workflow?  How can I "skip" to a certain stage or area of a workflow, bypassing the stuff that has already been done? 
    Correct me if Im wrong, but this seems to be where I need to look at bookmarks?  
    In my case, something similar to an order in a shopping cart is what would be the input type for the bigger workflow.  There are several different groups of users that have to set or add properties to associated objects (sql tables) related to the order
    before it can be marked "complete".
    How would you tie together a "wizard" type concept with workflow?  This seems to be the type of model we are considering.  A user takes an order a few steps, then another user from a different group/department adds some attributes and
    the workflow/wizard continues a few more levels down.  Finally, after all the rules have been matched and required data provided, the workflow can "finish".
    Hope my description seems clear.  

    My main unknown however, is how would a user interact with a workflow that was this high level.  Abstracting higher levels, means users dont need to interact with or execute the whole workflow each time, only a part of it.  So am I looking at pausing/resuming
    the bigger workflow?  How can I "skip" to a certain stage or area of a workflow, bypassing the stuff that has already been done? 
    If you want to skip to certain state or bypass what stuff that has been done, you need give logic judgment first. Bookmarks are the mechanism that enables an activity to passively wait for input without holding onto a workflow thread. When an activity signals
    that it is waiting for stimulus, it can create a bookmark. So bookmark is available for you during create this application.
    Here are some workflow applications, see:
    Workflow Application
    hope it helps.
    Regards,
    Angie
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • High-Level JTS/TopLink design question

    I've gone through the "using JTS with TopLink" docs, and it mostly makes sense. However, I still don't understand how TopLink "knows" when I call acquireUnitOfWork() whether or not I'm participating in a distributed 2PC transaction.
    Said another way:
    Let's say I've got an application based on TopLink (registering appropriate JTS stuff) that exposes an API that can be accessed remotely (RMI, SOAP, whatever).
    And, I've got another, separate application using a different persistence-layer technology (also supporting JTS) that also has an API.
    Now, I create a business method that uses the APIs from both of these applications, and I want them to participate in a single, distributed transaction.
    At a high level (source code is unnecessary), how does that work?
    Would the API need to support an ability to specifiy a TransactionContext or is this all handled behind the scenes by the 2 systems registering with the Transaction Service?
    If this is all handled through registration, how do these 2 systems know that these specific calls are all part of the same XA transaction?

    Nate,
    TopLink particiaptes in JTA/JTS transactions but dows not control them. When you configure TopLink to use the JTA/JTS services of the host application server you are deferring TX control to the J2EE container. TopLink will in this case register each acquired UnitOfWork in the current active TX from the container. The container will also ensure that the JDBC connection provided to TopLink is also bound by the active TX.
    In order to get 2PC you must register multiple resources into the same JTA TX. The TX processing during commit will then make the appropriate call backs to the underlying data source as well as the necessary call backs to listeners suchs as TopLink to have its SQL issued against the database.
    In short: The J2EE container manages the 2PC TX and TopLink is just a participant.
    Doug Clarke

  • Running a Sub-VI and monitoring data that is generated on a higher level VI

    Hi All, 
    This question must been there before, but I cannot find a suitable answer here on the forums....
    I have a 'top-level' VI that does a lot of things. I also have a sub VI that runs a frequency sweep on a piece of equipment. This is done with a for loop. 
    Problem: 
    I want to monitor/access the data that is generated in the for loop (See attached, the 3 wires within the green circle I want to monitor). 
    2 Questions:
    How can I access the data on the wires (within the loop) from a higher level VI?
    How can I then run this VI in a higher level VI while the higher level VI is continuing and not waiting for the sub-VI to complete?
    I tried using a Que but I cannot seem to get that working. 
    Any suggestions?
    Regards,
    Attachments:
    LV problem.PNG ‏44 KB

    The queue is a good way to move data from a running subVI to another VI.  Your problem is that if the subVI is inside a loop in the main VI, that loop in the main VI cannot iterate until the subVI completes. The solution: have the sub VI running in parallel - not inside - the loop.
    Look at the Producer/Consumer Design Patterns (at File >> New... >> VI >> From Template >> Frameworks >> Design Patterns >> Producer/Consumer.  This may be more than you need at the moment but will show how the parallel code process works.
    Lynn

  • Where can I find various high level examples of workflows being used

    I am about to start a project with TCS 3.5 and have been participating in the Adobe webinars to help learn components and specific techniques but what I am lacking is an understanding of various workflows I can model my project after or take bits from various sources. Why start with Framemaker in this workflow versus RoboHelp or even Word? Questions like this I think come from experience with the process and I am thinking that what I am getting myself into is a chessgame with all these pieces and don't want to paint myself into a corner by traveling down one route. I have seen this graphic:
    And this one:
    And this one:
    But they are too generic and do not contain enough information to really understand the descision making process one must go through on various projects.
    Can we have a series of webinars made, all with the underlining theme of defining a working process or workflow, by having guests describe how they have or are using this suite in real life on their own projects? One that might include a graphic showing the routes taken through the suite with reasons why?
    My project hopes to make a single source internal site that will tie together various 3D portable industrial coordinate metrology systems (hardware and software). It would be used as a dispersal site for help, communications between users and SME, OEM information, QA requirements, established processes, scripting snipet downloads, statistics, and training (including SOJT). Portable industrial metrology has 8 different softwares that are used and right now about 8 different instruments. These include laser trackers and radars, articulated arms, scanners, structered white and blue light to name a few. The softwares include Spatial Analyzer, Veriserf, CompIT, eMscon, AXYZ to a few there as well. I want to be able to participate and add content to an internal Sharpoint site, push content to users for stand-alone workstations, ePub, capture knowledge leaving the company through attrition, develop easy graphic rich job aid sheets, and aid in evaluations of emergent software and hardware. I would also like to leave the option open to use the finished product as a rosetta stone like translator between the software packages; doing this is the equivelent of doing this in these other software pacages for example.

    PDF is definately a format I want to include, to collaborate with other divisions and SME for one reason, but also for the ease in including 3D interactive target models with in it and portability. I plan on being able to provide individual PDFs that are very specific in their topics and to also use them to disperse user guides, cheat sheets or job aids... something the user may want to laminate on their own and keep with them for reference, printed out. Discussion in these sheets would be drasticly reduced to only the elements, relying heavely on bullet points or steps, usfull graphs, charts and tables... and of course illustrative images. I am thinking that these should be downloadable buttons to print on each topic section, not in a general apendix or such. They would hopefully be limited to one page, double sided 8x10.
    The cheet sheet would have a simplistic flow chart of how or where this specific topic fits in the bigger picture,
    The basic steps,
    Illustrations, equipment, setup
    Software settings for various situations in a table or chart,
    Typical result graph to judge with,
    Applicable QA, FAA regulation settings or concerns,
    Troubleshooting table,
    Topic SME contact info
    On the back, a screen shot infographic of software process
    The trouble here is that I have read that FM has a problem sometimes in succesfully transfering highly structured or formatted material to RoboHelp. Does this then mean that I would take it from FM straight to PDF?
    Our OEM material is very high level stuff... basicly for engineers and not shop floor users... but that is not to say they don't have some good material that could be useful. Our internal content is spread out across many different divisions and continents, with various ways of saying the same thing. This leads QA to interpret the information differently depending where the systems are put to work. We also have FAA requirements that need to be addressed and reminded to the user.
    Our company is starting to also see an exodus of the most knowledagble of the users through retirement. Capturing the knowledge and soft skill packages they have developed working here for 20-30 years is something I am really struggling with. I have only come up with two ideas so far:
    Internal User Web based Forum
    Interviews (some SMEs do not want to make the effort in transfering knowledge by participating in anything if it requires an effort they don't see of benefit to themseleves), to get video, audio or transcription records

  • Phase out settings at a higher level such as brand or major customer

    Have any of you ever setup phase out assignments at a higher level than product and had it work correctly?  For example, we want to phase out a brand for a major customer.  In other words, a customer is dropping a brand and we don't want statistical forecast generated for that brand customer combination any longer.  I am able to setup the fields in phase out lifecycle settings for product, brand, major account but when i enter the brand and major account I am still getting forecast generated.  It appears to stop for some products within the brand but not all.  Another example is if a customer quits ordering from us I want to setup the major customer to phase out so no forecast is generated.
    If you have done this successfully please let me know.  Or if you would handle these situations in a different manner other than phase out please let me know.  We can do historical adjustments each period but that is a lot of maintenance to do after each period before statistical forecast is generated.
    Thanks
    Steve

    Hi Stephen,
    Life cycle planning works only at the detail level (each CVC), the option of aggregate planning is helpful if you want to phase in or out a certain CVC when you are forecasting at the aggregate level.
    One option is that you should have all products in the "profile assignment for life cycle " section which falls under that brand and customer. you can maintain a file and then automate the upload process in to the "assignment"
    or
    you can try to use the copy functionality in the realignment (/SAPAPO/RLGCOPY) , where you maintain copy factor as NIL and when the stat fcst is generated, you can use this as the next step to zero out -but you would need to maintain them manually.
    or
    easiest and safest wayy would be create a selection for those combination and do not include them in the planning job for stat fcst.
    or
    you can build a customised program to access the PA, PB, Data view and input the selection to zero out the stat fcst KF for that particular selection after the stat fcst run. here you would need to check if the diaggregated values are good enough.
    hope it helps.

  • Errors in the high-level relational engine. A connection could not be made to the data source with the DataSourceID of '

    When I deploy the cube which is sitting on my PC (local) the following 4 errors come up:
    Error 1 The datasource , 'AdventureWorksDW', contains an ImpersonationMode that that is not supported for processing operations.  0 0 
    Error 2 Errors in the high-level relational engine. A connection could not be made to the data source with the DataSourceID of 'Adventure Works DW', Name of 'AdventureWorksDW'.  0 0 
    Error 3 Errors in the OLAP storage engine: An error occurred while the dimension, with the ID of 'Customer', Name of 'Customer' was being processed.  0 0 
    Error 4 Errors in the OLAP storage engine: An error occurred while the 'Customer Alternate Key' attribute of the 'Customer' dimension from the 'Analysis Services Tutorial' database was being processed.  0 0 

    Sorry hit the wrong button there. That is not entire solution and setting it to default would work when using a single box and not in a distributed application solution. If you are creating the analysis database manually or using the wizard then you can
    set the impersonation to your heart content as long as the right permission has been set on the analysis server.
    In my case I was using MS Project Server 2010 to create the database in the OLAP configuration section. The situation is that the underlying build script has been configured to use the default setting which is the SQL Service account and this account does
    not have permission in Project Server I believe.
    Changing the account to match the Project service account allowed for a successful build \ creation of the database. My verdict is that this is a bug in Project Server because it needs to include the option to choose impersonation when creating the Database
    this way it will not use the default which led to my error in the first place. I do not think there is a one fix for all in relations to this problem it is an environment by environment issue and should be resolved as such. But the idea around fixing it is
    if you are using the SQL Analysis server service account as the account creating the database and cubes then default or service account is fine. If you are using a custom account then set that custom account in the impersonation details after you have granted
    it SQL analysis administrator role. You can remove that role after the DB is created and harden it by creating a role with administrative permissions.
    Hope this helps.

  • Error: Maintain settlement rule of the sender for a higher level WBS

    Hi,
    I dont want to maintain the settlement rule for a higher level WBS. How can i configure this in such a way that i dont get the following error:" Maintain settlement rule of the sender" while doing CJ88. Maintaing a separate Settlement profile for a higher level WBS is an option but we are looking if something else could be done The problem is that there are no actuals booked against, say, level 2 WBS but when i execute CJ88, i get the aforesaid error. How can i ensure that only the lowest level WBS ask for the settlement rule and not the levels above it. I have already removed the Investment profile from the higher level WBS but still getting the same error.
    Regards,
    DPil

    Hi,
    It is a type Capex WBS and Biling element is not checked. In fact i get a warning while doing the settlement: WBS is neither a billing element nor an account assignment element.
    Diagnosis
    WBS element  is not indicated as either an account assignment element or as a billing element in the master record.
    System Response
    The WBS element cannot be assigned to an account.
    Procedure
    Correct your entries or add the missing indicator to the master record for the WBS element.
    But this is just a warning. On pressing enter i get the error :Maintain settlement rule of the sender "

  • High level & low level Desing in OBIEE

    Hi Gurus/Experts,
    I am new to in OBIEE and please let me know about the High Level and Low level design of obiee projects.
    Really its helpful for me to develope my carrier skills.
    Thanks in advance,
    Sriram

    Hi,
    OBIEE 11g Basic Security Guide from Deliver BI
    http://www.box.net/shared/5ef1alb2sp
    http://www.rittmanmead.com/2008/04/migration-obiee-projects-between-dev-and-prod-environments/
    http://obiee101.blogspot.com/2009/07/obiee-how-to-get-started.html
    http://www.scribd.com/doc/60264784/OBIEE11g-Logical-Table-Souorce
    tutorial from the Oracle By Example range
    http://st-curriculum.oracle.com/obe/fmw/bi/bi1113/createanalysis/ps.htm
    also refer the blow one for installation and migration and essential ..basics..
    OBIEE10 repository and webcatlouge in obiee 11g
    http://www.rittmanmead.com/2010/08/oracle-bi-ee-11g-upgrading-from-bi-ee-10g-repository-web-catalog/
    new features OBIEE 11g is having
    http://obieeelegant.blogspot.com/2011/10/obiee11g-features.html
    There are the full new features about OBIEE 11g.
    Top5 New Features in Oracle Business Intelligence Security
    http://oracleintelligence.blogspot.com/2010/10/top5-new-features-in-oracle-business.html
    Top5 New Features for Oracle Business Intelligence System Administrators
    http://oracleintelligence.blogspot.com/2010/10/top5-new-features-for-oracle-business.html
    Top10 New Features for Oracle Business Intelligence Users
    http://oracleintelligence.blogspot.com/2010/10/top10-new-features-for-oracle-business.html
    List of Bug Fixes Included In OBIEE 11.1.1.5.0
    http://obieeelegant.blogspot.com/2011/11/list-of-bug-fixes-included-in-obiee.html
    OBIEE 11g (11.1.1.5.0) Software Only Installation on Windows Server 2008/2003/XP 64 bit SP1 &SP2
    http://obieeelegant.blogspot.com/2011/09/obiee-11g-111150-software-only.html
    Migrating Application Roles from Dev to UAT server and Production server.
    To move Application Roles, please kindly review the following information:
    Oracle Fusion Middleware Application Security Guide 11g Release 1 (11.1.1)
    7.3.2 Migrating Policies with the Command migrateSecurityStorehttp://download.oracle.com/docs/cd/E14571_01/core.1111/e10043/cfgauthr.htm#JISEC2929
    http://www.rittmanmead.com/2010/10/obiee-11gr1-security-explained-working-with-the-default-security-configuration/
    http://www.rittmanmead.com/2011/04/oracle-bi-ee-11g-migrating-security-credential-store-part-3/
    The same guide can be used...But some fetures are changed in it.Here is the RPD 11g step-by-step guide
    http://www.oracle.com/webfolder/technetwork/tutorials/obe/fmw/bi/bi11115/biadmin11g_02/biadmin11g.htm
    Refer:
    Re: OBIEE 11G Beginners guide
    Thanks
    Deva

  • How to perform high-level planning without doing full costctr-level rollup

    I am wondering what my options are for this effort - we want to start our 2009 budget season by first preparing a high-level 2009 plan. To do the high-level plan we'd like to avoid building and interfacing workbooks at the cost center level; instead, to get started, we want to look at groups of cost centers that represent the different lines of business of the Bank (retail, credit, etc).
    Here's my question:is it possible to build planning workbooks not at the cost center level, but at a cost center rollup level? That is, we'd like to avoid having 10 workbooks at the cost center level that make up the rollup "Mortgage Lending"; instead we want one workbook for Mortgage lending that includes data for all 10 of those CC's.
    Please let me know if you have any ideas as to how to do so, thanks, W

    I am wondering what my options are for this effort - we want to start our 2009 budget season by first preparing a high-level 2009 plan. To do the high-level plan we'd like to avoid building and interfacing workbooks at the cost center level; instead, to get started, we want to look at groups of cost centers that represent the different lines of business of the Bank (retail, credit, etc).
    Here's my question:is it possible to build planning workbooks not at the cost center level, but at a cost center rollup level? That is, we'd like to avoid having 10 workbooks at the cost center level that make up the rollup "Mortgage Lending"; instead we want one workbook for Mortgage lending that includes data for all 10 of those CC's.
    Please let me know if you have any ideas as to how to do so, thanks, W

  • Why does OWB 9.2 generate UK's on higher levels of a dimension?

    When you specify levels in a dimension, OWB 9.2 generates unique key constraints in the table properties for every level, but only the UK on the lowest level is visible in the configuration properties. Why then are these higher level UK's generated? Is this a half baked attempt to implement the possiblility to generate a snow flake model in OWB?
    Jaap.

    Piotr, Roald and others,
    This is indeed a topic we spend a lot of our time on these past months. We are addressing this as (in my old days I had the same problem as a consultant) we know that this is a common problem.
    So the solution is one that goes in 2 directions:
    - Snowflake support
    - Advanced dimension data loading
    Snowflake is obvious, may not be desired for various reasons but we will start supporting this and loading data for it in mapping.
    If you want a star table, you will know that a completely flattened table with day at the lowest level will not be able to get you a unique entry for month. So what people tend to do is one of the following:
    - Proclaim the first of the month the Month entry point (this stays closest to the star table and simply relies on semantics on both ETL and query side).
    - Create extra day level entries which simbolize the month, so you have a day level with extra entries
    - Create views, extra tables etc to cover the extra data
    - Create a data set within the tables that solves the key problem
    We have opted for the last one. What you need to do for this is a set of records that uniquely identify any record in any level. Then you add a key which links to the dimension at the same point (a dimension key), so all facts always use this surrogate key to link (makes life in query tools easier).
    For a time dimension you will have a set of day records with their months etc in them (the regular star). Then you add a set of records with NULL in the day having months and up. And you go up the hierarchy. For this we will have the ETL logic (in other words you as a designer do not worry about this!). On the query tool you must be a little cautious on counts but this is doable and minor.
    As you can see none of the solutions are completely transparent, but we believe this is one that solves a lot of problems and gives you the best of all worlds. We will also support the same data structure in the OLAP dimensions for the database as well in the relational dimension. NOTE that there are some disclaimers with this as we are doing software here...
    In principal however we will solve your problem.
    Hope this explains some of our plans in this area.
    Jean-Pierre

Maybe you are looking for

  • After upgrading to Lion Time Capsule solid yellow

    Hi After upgrading to lion I am experiencing a lot of problems with time caspule 1. backup fail: it starts and than fails (error message and yellow warning triangle) 2. unable to connect to TC disk, neither from airport utility, nor from Finder 3. Ai

  • CRM 5.0 ATP check to ECC not including pick / pack time in ECC 6.0

    Been searching SDN & OSS, our ATP check in 5.0 is reading the proper inventory in ECC but when no inventory is available the date being returned is the total replenishment leadtime in the ECC material master. (good) When the order is entered in ECC i

  • Agentry HTTP-XML Step Bad Request Error

    Hi, I try to get data via HTTP-XML connection on Agentry. 1)  I defined HTTPXML Sytem Connection then added Validate User Request for GET method with URL: http://www.thomas-bayer.com/sqlrest/PRODUCT/      I did not defined Request Argumanets because

  • Issue with Service PO Process in Production Server

    Hi, We are doing the Process of Service PO. First create Service Master through AC01 with Material Group 603010. Now during Service PO, by mistake we insert Material Group 605010 inplace of 603010 in PO line item...After that GR has been done... Till

  • Javaw.exe throws Disk A: is not ready message

    My application worked fine with JWS 1.0.1 until I started using JFileChooser in my application. When I am calling a constructor JFileChooser("C:/TEMP") JWS's javaw.exe throws the warning "Disk A is not ready. Retry, Abort, Ignore" Then if I press "Ig