Accessing Data Warehouse with HTML DB

I have a test data warehouse database 10g comprising of seven dimension tables and one fact table. When I access one table at a time, the query runs fine, but when I join two dimension tables or more to the fact table, the result set comes out wrong. The performance is also very poor. Is HTML DB not capable of properly accessing a data warehouse data?
Here is the query I'm having problem with:
SELECT p.prod_name, s.store_name, pr.week, sl.dollars
FROM sales sl, product p, period pr, store s
WHERE p.prodkey = sl.prodkey
AND pr.perkey = sl.perkey
AND p.prod_name LIKE 'Assam Gold%'
OR p.prod_name LIKE 'Earl%'
AND s.store_name LIKE 'Instant%'
AND pr.month = 'NOV'
AND pr.year = 2003
ORDER BY p.prod_name, sl.dollars DESC
Your input would be appreciated.

I doubt this was intentional, but you are not joining the store table to anything. You do filter the rows from that table with the AND s.store_name LIKE 'Instant%' predicate, but it is not joined to any of the other 3 tables. Your query will essentially return the number of rows from the other 3 tables multiplied by the number of rows returned from store. SYou might think about grouping some of your predicates for readability and possibly for correct logic.SELECT p.prod_name, s.store_name, pr.week, sl.dollars
  FROM sales sl, product p, period pr, store s
WHERE p.prodkey = sl.prodkey
   AND pr.perkey = sl.perkey
   -- Add missing predicate here
   -- AND s.something = sl,p, or pr .something
   -- end missing predicate
   AND (p.prod_name LIKE 'Assam Gold%'
        OR
        p.prod_name LIKE 'Earl%')
   AND s.store_name LIKE 'Instant%'
   AND pr.month = 'NOV'
   AND pr.year = 2003
ORDER BY p.prod_name, sl.dollars DESCHope this helps,
Tyler

Similar Messages

  • How to create and use Data Vault with HTML/JavaScript (SAPUI5) for Mobile Apps?

    Hello,
    I am creating a demo Enterprise Mobile App, for cross platform and I want to use the Data Vault.
    I am not able to figure out how to implement the Data Vault with HTML/JavaScript i.e. SAPUI5 for cross platform mobile apps I have a tutorial, but it is for Android based apps. Where as I want it for iOS as well. I guess, in this case, implementing the Data Vault using SAPUI5 over HTML and JavaScript would be  better.
    If anyone has any links or sample code to implement Data vault in HTML/JavaScript (SAPUI5) or specific for iOS apps, it would be great.
    Thank you.

    Hi,
    This is the "SAP Mobile Documents" community, so this seems to be the wrong place for asking your question related to Data Vault / SAPUI5.
    Maybe you should post your question here: http://scn.sap.com/community/developer-center/front-end.
    best regards,
    Ingo

  • Accessing many databases with HTML DB from a single Apache Server

    Hello;
    I am planning the install of 10g and HTML DB on a HPUX server with many databases running out of many Oracle Homes 9.2.0, 9.0.1 and 8.1.7. I will be upgrading a number of the existing databases to 10g and want to access them with HTML DB.
    My question is how can I install HTML DB on many databases and use a single Apache Server to access them?

    Where would I find the DADs? a default installation of htmldb puts the modplsql DAD, "pls/htmldb", in a marvel.conf file located in $OHS_HOME/Apache/modplsql/conf/marvel.conf
    What would the URLs look like? not sure what you're asking. if you're asking about how the change in DAD's would affect the URLs you're calling, though, you should consider looking over the OHS (oracle http server) doc some more before making these types of customizations.
    And how would I install the HTML DB in the different databases?
    Would I use the installer or scripts?the only supported way to install htmldb is via the 10g Companion CD. just use that installer to point to your different db's.
    regards,
    raj

  • Tungsten T3 - need to access DAT file with addresses - Please Help

    I have an old Palm Tungsten T3. The desktop computer with the synced address data has died and the T3 has lost power (so no addresses anymore).
    I have a .DAT file backup with all my addresses. Can anyone advise how I can get access to these.
    Reinstalling Palm Desktop 4.1.4e did not help as it doesn't allow me to import .DAT files.
    Can I import this file into any other app or open up in Excel etc?
    Any advice would be much appreciated.
    Thanks in advance.
    Post relates to: Tungsten T3
    This question was solved.
    View Solution.

    You have to physically copy the .dat file into the User directory over the existing address.dat file.  Find the one with your hotsync name abbreviated (i.e. - C:/Program Files/Palm (or PalmOne)/ your Hotsync name /address) and copy the .dat file there.
    Be sure you copy the file and keep another one safe in case something goes wrong.  The existing blank address.dat file will be only about 2k in size.
    Good luck!
    WyreNut
    I am a Volunteer here, not employed by HP.
    You too can become an HP Expert! Details HERE!
    If my post has helped you, click the Kudos Thumbs up!
    If it solved your issue, Click the "Accept as Solution" button so others can benefit from the question you asked!

  • Unable to Access Data Form with Users Provisioned as 'Planner'

    Hello,
    I am currently experiencing an issue that I can't seem to resolve. I was hoping someone else might have some insight and/or have come across this issue.
    I created two data forms. Users have been assigned permissions to the dimensions included in the data form. I have also assigned users access to the data form. The model is a single currency model.
    Upon logging into the application, the user is unable to open the data form unless he/she is provisioned as an 'Administrator'. Provisioning the user as a 'Planner' or 'Interactive User' does not allow the user to open the data form. The following message appears:
    *"Security and/or filtering has resulted in a required dimension not being represented on this data form."*
    My data form uses the following dimensions:
    POV: ELEMENT
    PAGE: SCENARIO, YEAR, VERSION, ENTITY
    ROWS: ACCOUNT
    COLUMNS: PERIOD
    We are currently implementing Hyperion Planning 11.1.2.0. The operating system is Windows Server 2008 64-bit R2.
    Any advice/suggestions on how to resolve this would be greatly appreciated. Thank you.
    Corey

    I went ahead a created a data form that just used level 0 members and confirmed that the user had access to the members defined in the form. Still no luck.
    I also checked to make sure the filters existed in Essbase. Here is what I found in Essbase:
    Access: None
    Member Specification: @IDES("Account"),@IDES("Scenario"),@IDES("Version"),@IDES("Entity"),@IDES("NTTA_ELEMENT")
    I have no issues accessing the data through the Essbase Add-in.
    I did notice some additional items in my Planning setup:
    1. Task>Manage Process - Sub-Status states 'Failed' when I attempt to validate
    2. Task>Manage Process - Status changes from 'Not Started' to 'Under Review' when I start the cycle.
    3. Dimension Security - Users are attached to a group that has write access at the Descendent(Inclusive) level.
    4. In the Account, Department, Scenario and Version dimensions I also setup the user to have write access at the member level.
    As for my earlier post with the data forms, that was just my way of showing the dimensions used. I do have two data forms, 1) with top level and descendants/children and 2) with level 0 members only.
    I checked the logs and there isn't really much info there. I have a case open with Oracle and a webex scheduled for tomorrow.
    Thoughts?
    Thanks
    Corey

  • Create schemas for Reporting Data Warehouse with Oracle XE

    There is the possibility to import dw house and loader with the database Oracle XE?
    I receive this error ORA-00439: feature not enabled: Advanced replication in CIM log.
    I saw that the database, we do not have the feature 'Advanced replication'.
    SQL> select * from v $ option where parameter = 'Advanced replication';
    PARAMETER
    VALUE
    advanced replication
    FALSE
    Log CIM:
    **** info Mon Feb 23 14:16:00 BRT 2015 1424711760686 atg.cim.database.dbsetup.CimDBJobManager Top level module list for Datasource Reporting DataWarehouse: DCS.DW,ARF.DW.base,ARF.DW.InternalUsers,Store.Storefront
    **** info Mon Feb 23 14:16:05 BRT 2015 1424711765012 atg.cim.database.dbsetup.CimDBJobManager 0 of 0 imports not previously run.
    **** info Mon Feb 23 14:16:05 BRT 2015 1424711765192 atg.cim.database.dbsetup.CimDBJobManager Top level module list for Datasource Reporting Loader: DafEar.Admin,DCS.DW,DCS.PublishingAgent,ARF.base,Store.EStore,Store.EStore.International
    **** info Mon Feb 23 14:16:05 BRT 2015 1424711765733 atg.cim.database.dbsetup.CimDBJobManager 1 of 1 imports not previously run.
    **** info Mon Feb 23 14:16:05 BRT 2015 1424711765953 atg.cim.database.dbsetup.CimDBJobManager Top level module list for Datasource Publishing: DCS-UI.Versioned,BIZUI,PubPortlet,DafEar.Admin,DCS-UI.SiteAdmin.Versioned,SiteAdmin.Versioned,DCS.Versioned,DCS-UI,Store.EStore.Versioned,Store.Storefront,DAF.Endeca.Index.Versioned,DCS.Endeca.Index.Versioned,ARF.base,DCS.Endeca.Index.SKUIndexing,Store.EStore.International.Versioned,Store.Mobile,Store.Mobile.Versioned,Store.Endeca.International,Store.KnowledgeBase.International,Portal.paf,Store.Storefront
    **** info Mon Feb 23 14:16:11 BRT 2015 1424711771561 atg.cim.database.dbsetup.CimDBJobManager 65 of 65 imports not previously run.
    **** info Mon Feb 23 14:16:11 BRT 2015 1424711771722 atg.cim.database.dbsetup.CimDBJobManager Top level module list for Datasource Production Core: Store.EStore.International,DafEar.Admin,DPS,DSS,DCS.PublishingAgent,DCS.AbandonedOrderServices,DAF.Endeca.Index,DCS.Endeca.Index,Store.Endeca.Index,DAF.Endeca.Assembler,ARF.base,PublishingAgent,DCS.Endeca.Index.SKUIndexing,Store.Storefront,Store.EStore.International,Store.Recommendations,Store.Mobile,Store.Endeca.International,Store.Fluoroscope,Store.KnowledgeBase.International,Store.Mobile.Recommendations,Store.Mobile.International,Store.EStore,Store.Recommendations.International
    **** info Mon Feb 23 14:16:12 BRT 2015 1424711772473 atg.cim.database.dbsetup.CimDBJobManager 30 of 30 imports not previously run.
    **** info Mon Feb 23 14:16:19 BRT 2015 1424711779573 atg.cim.database.dbsetup.CimDBJobManager Creating Schema for Datasource Reporting DataWarehouse
    **** info Mon Feb 23 14:16:19 BRT 2015 1424711779653 atg.cim.database.dbsetup.CimDBJobManager Top level module list for Datasource Reporting DataWarehouse: DCS.DW,ARF.DW.base,ARF.DW.InternalUsers,Store.Storefront
    **** info Mon Feb 23 14:16:19 BRT 2015 1424711779993 atg.cim.database.dbsetup.CimDBJobManager Create DatabaseTask for Module ARF.DW.base, sql/db_components/oracle/arf_ddl.sql
    **** info Mon Feb 23 14:16:19 BRT 2015 1424711779993 atg.cim.database.dbsetup.CimDBJobManager Create DatabaseTask for Module ARF.DW.base, sql/db_components/oracle/arf_view_ddl.sql
    **** info Mon Feb 23 14:16:19 BRT 2015 1424711779993 atg.cim.database.dbsetup.CimDBJobManager Create DatabaseTask for Module ARF.DW.base, sql/db_components/oracle/arf_init.sql
    **** info Mon Feb 23 14:16:19 BRT 2015 1424711779993 atg.cim.database.dbsetup.CimDBJobManager Create DatabaseTask for Module DCS.DW, sql/db_components/oracle/arf_dcs_ddl.sql
    **** info Mon Feb 23 14:16:19 BRT 2015 1424711779993 atg.cim.database.dbsetup.CimDBJobManager Create DatabaseTask for Module DCS.DW, sql/db_components/oracle/arf_dcs_view_ddl.sql **** info Mon Feb 23 14:16:19 BRT 2015 1424711779993 atg.cim.database.dbsetup.CimDBJobManager Create DatabaseTask for Module DCS.DW, sql/db_components/oracle/arf_dcs_init.sql
    **** info Mon Feb 23 14:16:21 BRT 2015 1424711781085 atg.cim.database.dbsetup.CimDBJobManager Found 2 of 6 previously unrun tasks for Datasource Reporting DataWarehouse
    **** info Mon Feb 23 14:16:21 BRT 2015 1424711781085 atg.cim.database.dbsetup.CimDBJobManager 1 ARF.DW.base : sql/db_components/oracle/arf_view_ddl.sql
    **** info Mon Feb 23 14:16:21 BRT 2015 1424711781085 atg.cim.database.dbsetup.CimDBJobManager 2 DCS.DW : sql/db_components/oracle/arf_dcs_view_ddl.sql
    **** info Mon Feb 23 14:16:21 BRT 2015 1424711781085 /atg/dynamo/dbsetup/job/DatabaseJobManager Starting database setup job 1424711781085.
    **** Error Mon Feb 23 14:16:21 BRT 2015 1424711781516 /atg/dynamo/dbsetup/database/DatabaseOperationManager --- java.sql.SQLException: ORA-00439: feature not enabled: Advanced replication
    is there a workaround?

    Hi
    We haven't tested and certified with Oracle XE internally
    You need to use an  Oracle Enterprise Edition for Advanced Replication
    What version of Oracle Commerce are you installing, I cannot tell from the log snippet you have posted
    ++++
    Thanks
    Gareth
    Please mark any update as "Correct Answer" or "Helpful Answer" if that update helps/answers your question, so that others can identify the Correct/helpful update between many updates.

  • Syntax for WriterLoginName in Data Warehouse DB

    Hello
    I'm having a few issues with our management servers writing to the Data Warehouse DB. I've checked the 'Management Group' table and can see the WriterLoginName is set to
    DOMAIN\sv-scom-dw - however, i'm just querying whether that field should read
    sv-scom-dw
    The account is in fact a domain account. It's listed as the 'Data Warehouse SQL Account' & 'Data Warehouse Action Account' (under Administration > Run As configuration > Accounts). 
    We have two entries in the database security (rights over OperationsMangerDW), one as DOMAIN\sv-scom-dw & a local SQL login called sv-scom-dw. Both accounts have the following permissions: apm_datareader, apm_datawriter, db_datareader, db_owner, OpsMgrReader,
    OpsMgrWriter, public.
    We're a SCOM 2012 R2 environment. All servers are 2012 R2, SQL is also 2012 standard. 
    Anyone faced a similar issue before? I'm seeing a lot of alerts in the Monitoring section for the Data Warehouse. One in particular:
    Data Warehouse failed to discover performance standard data set. Failed to enumerate (discover) Data Warehouse objects and relationships among them. The operation will be retried.
    Exception 'SqlException': Management Group with id ''5F201AB2-4B10-7FCC-C716-B2361102248D'' is not allowed to access Data Warehouse under login ''sv-scom-dw''
    One or more workflows were affected by this.
    Workflow name: Microsoft.SystemCenter.DataWarehouse.Discovery.StandardDataSet
    Instance name: Performance data set
    Instance ID: {B81C47FB-A80D-0FE5-A8DB-DC4544FC8DA6}
    Management group: ******
    As you can see from the alert the account referenced is 'sv-scom-dw' and not 'DOMAIN\sv-scom-dw'. Which is why I originally asked, should the field in the management table be updated?
    Thanks, David.

    Hi guys.
    Thanks for the responses, I shall provide an event  ID shortly. In response to Mai, I've followed the link you've posted and I'm now checking the 'data source and related settings', so i've gone to http://localhost/reports on the Warehouse server (which
    also hosts the reporting), and i've got the following error:
    The report server cannot decrypt the symmetric key that is used to access sensitive or encrypted data in a report server database. You must either restore a backup key or delete all encrypted content. (rsReportServerDisabled)
    Get
    Online Help
    Keyset does not exist (Exception from HRESULT: 0x80090016)
    Have you come across this before?

  • Data Warehouse errors, SQL Server Agent Stopped

    I just installed SCE 2010 on a 2k8 server and went through the config wizard.  Now it is showing Data Warehouse alerts, including:
    Data Warehouse failed to discover object health state standard data set. Failed to enumerate (discover) Data Warehouse objects and relationships among them. The operation will be retried.
    Exception 'SqlException': Management Group with id '6FDBE66E-E35D-BCB8-0FA1-5947E45FE643' is not allowed to access Data Warehouse under login 'domain\user'
    One or more workflows were affected by this.
    Workflow name: Microsoft.SystemCenter.DataWarehouse.Discovery.StandardDataSet
    Instance name: State data set
    Instance ID: {217CFE24-F6F6-0DFF-9EA9-64C62895EE29}
    Management group: SCESERVER_MG
    I let the SCE install SQL 2008 Express.  When I checked on the SQL Services, SQL Server Agent Service is stopped and will not start.  Not sure if this is the cause of the above issue.  I also can't log in to the SQL Server Management Studio.
    How can I fix this?  What other information do you need from me.

    Danny01 - tried the domain admin rights on the service with no difference - put it back
    Radhey Shah - I checked the admin account and it is the same account that I set the SQL up with.  How do I verify this account has admin rights on the SQL DB?
    The error that comes up when I try to run the SQL Server Agent Service is...
    "The Sql Server Agent (Essentials) service on SCESERVER
    started and then stopped.  Some services stop automatically if
    they are not in use by other serives or programs."
    Also, when I look at the alerts within SCE, here is the current alert..
    OleDb Module encountered a failure 0x80004005 during execution and will post it as output data item. Unspecified error
    : Cannot open database "OperationsManagerDW" requested by the login. The login failed.
    Workflow name: Microsoft.Windows.Client.Vista.ComputerGroup.DiskTrendsComputer
    Instance name: Microsoft System Center Data Warehouse
    Instance ID: {16781F33-F72D-033C-1DF4-65A2AFF32CA3}
    Management group: SCESERVER_MG

  • Data Warehouse issues

    Hi there.
    I'll try and be to the point as I can. I undertook an upgrade from SCOM 2012 SP1 to R2, this was successful. However after a few days had past, I noticed that the log file for the DW DB was growing rapidly, to the point where we hit a real issue. The drive
    was a comfortable 50GB drive, but in the end had to be increased to just short of 500GB !! Long story short, I wasn't allowed to get a consultant in to help diagnose the issue, therefore, I was consulting Google, Technet articles etc.
    The end decision was to DELETE the Data Warehouse DB & log file and create them again from scratch. I've done this, and used the following instructions on how to do so (http://www.3packetsofcrisps.com/2014/04/reset-scom-2012-data-warehouse-database.html).
    I have data going into the DW DB, or rather did up until midnight, so had around 3.5 hrs worth. 
    After checking the Operations Manager event log on one of the Management Servers, I can see numerous errors, these include:2115, 4506 and occasionally 8000. 
    The event 4506 states: "Data was dropped due to too much outstanding data in rule "Rule Name" running for instance "instance name" with id "ID name" in management group "xxx". 
    After some consulting Google, I read that this might be expected because a lot of data is trying to be written to the DW DB, however, that doesn't explain why it was writing and for some reason randomly stopped. 
    A quick summary of our environment:
    2x Management Servers running Win 2012 Std x64
    1x Operations Database server running Win 2012 Std x64 & SQL Server 2012 Std x64
    1x Data Warehosue DB server running Win 2012 Std x64 & SQL Server 2012 Std x64
    I've checked account permissions using http://technet.microsoft.com/en-gb/library/hh457003.aspx
    If you need any more information, please say, I've tried to keep the post as small as possible.
    Dave.

    Hi 
    Thanks for responding. It's been over a day so far, however, I've only just flushed the health service state & cache for the servers I'm monitoring. Would you suggest I do this for the Management Servers as well?
    In regards to alerts in the SCOM console, I am getting active alerts. I've just checked and some were 49 minutes ago. 
    I've also taken another look at the DW DB (via Windows Explorer), and I can see the last modified date has changed to 09/09/2014 20:48. So once again, something has happened but it appears to be random..
    This might need to be a separate post, but I've also got issues with my Reporting. If I go to the 'Reporting Server URL', I get the following message: "The version of the report server database is either in a format that is not valid, or it cannot
    be read. The found version is ''. The expected version is '162'. (rsInvalidReportServerDatabase)"
    Any ideas?
    Dave

  • How i create oracle9i data Warehouse

    hi master
    sir i am use oracle 9i with devloper 6i
    sir how i create data Warehouse with my 9i data base
    please give me step
    what software i install
    how i use portal
    how i create datamart
    how create cub
    how i connect my oracle 9i with Warehouse
    how i show my portal on net
    how i import my oracle 9i data in Warehouse
    how i create multidimation report for net becuase my other office in other
    city
    how i import diffrence type data from diffrence type of database
    please give me idea how i start and how i get above result
    thank's
    aamir

    You need to have the enterprise version of Oracle DB 9.2.0.2 in order to have DM4J to work fully. It will run on Oracle DB 9.2.0.1 but you would need to install a JMS patch to have the components run successfully. This is better done by just upgrading to 9.2.0.2.
    If you have the correct DB then you need to establish a connection to the db using the "ODM" account and password. You can not access data mining in the db other then through this account.
    Additionally, all ODM_MTR tables are granted to ODM for access out of the box. If you want more data accessible to ODM you will need to grant these individually.

  • Table and Index compression in data warehouse - thoughts?

    Hi,
    We have a data warehouse with large fact tables and materialized views of this data.
    Approx 3 million inserts per day week-ends about 12 million.
    The fact tables we have expected to have 200 million, and couple with 1-3 billion.
    Tables partitioned and have bitmap indexes.
    Just wondered what thoughts were about compressing large fact tables and mviews both from point of view of ETL into them and reporting from them afterwards.
    I take it, can compress/uncompress accordingly without any problem?
    Many Thanks

    After compression, most SELECT statements would not get slower. Actually, many can get faster due to reduced IO and buffer needs.
    The situation with DMLs is more complex. It depends on the exact compression options (basic or advanced) and the DML (INSERT,UPDATE, direct load,..),but generally DML are negatively affected by compression.
    In a Data Warehouses (DWs), it is usually quite beneficial to compress partitions or tables that contain data that is not supposed to be modified (read only or read mostly). Please note that in many cases you do not have to compress while you are loading the data – you can do that later.
    You can also consider compressing some of your B-tree indexes (if you use them in your DW system).
    Iordan Iotzov
    http://iiotzov.wordpress.com/

  • Why do we need SSIS and star schema of Data Warehouse?

    If SSAS in MOLAP mode stores data, what is the application of SSIS and why do we need a Data Warehouse and the ETL process of SSIS?
    I have a SQL Server OLTP database. I am using SSIS to transfer my SQL Server data from OLTP database to a Data Warehouse database that contains fact and dimension tables.
    After that I want to create cubes using SSAS form Data Warehouse data.
    I know that MOLAP stores data. Do I need any Data warehouse with Fact and Dimension tables?
    Is not it better to avoid creating Data warehouse and create cubes directly from OLTP database?

    Another thing to note is data stored in transactional system may not always be in end user consumable format for ex. we may use bit fields/flags to represent some details in OLTP as storage required ius minimum but presenting them as is would not make any
    sense to user as they would not know what each bit value represents. In such cases we apply some transformations and convert data into useful information for users to understand. This is also in the warehouse so that information in warehouse can directly be
    used for reporting. Also in many cases the report will merge data from multiple source systems so merging it on the fly in report would be tedious and would have hit on report server. In comparison bringing them onto common layer (warehouse) and prebuilding
    aggregates would be benefitial for the report performance.
    I think (not sure) we join tables in SSAS queries and calculate aggregations in it.
    I think SSAS stores these values and joined tables and we do not need to evaluates those values again and this behavior is like a Data Warehouse.
    Is not it?
    So if I do not need historical data, Can I avoid creating Data Warehouse?
    On the backend SSAS uses queries only to extract the data
    B/w I was not explaining on SSAS. I was explaining on what happens inside datawarehouse  which is a relational database by itself. SSAS is used to built cube (OLAP structures) on top of datawarehouse. star schema is easier for defining relationships
    and buidling aggregations inside SSAS as its simple and requires minimal lookups to be performed. Also data would be held at lowest granularity level which can easily be aggregated to required levels inside OLAP cubes. Cube processing is very resource
    intensive and using OLTP system would really have a huge impact on processing performance as its nnot denormalized and also doing tranformation etc on the fly adds up to complexity. Precreating a layer (data warehouse) having data in required format would
    make cube processing easier and simpler as it has to just cross join tables and aggregate data based on relationships defined and level needed inside the cube.
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • How to install and integrate oracle 10g and 11g Warehouse with EBS???

    Hi,
    We are using EBS r12 (12.0.6) and database 12.2.0.3.
    Please give us the Doc Id's for installing and confihuration oracle Data Warehouse with EBS r12?
    thx

    Hi;
    We are using EBS r12 (12.0.6) and database 12.2.0.3.you mean db version is 10.2.0.3 right? Coz there is no version like start 12 yet!
    Please give us the Doc Id's for installing and confihuration oracle Data Warehouse with EBS r12?Check below:
    Oracle E-Business Intelligence Embedded Data Warehouse Implementation Guide Release 4.1 [ID 246156.1]
    Embedded Data Warehouse White Papers [ID 149406.1]
    Embedded Data Warehouse Library [ID 149404.1]
    Regard
    Helios

  • Data Warehouse

    Hi
    I want to setup a very small data warehouse with one table as a sample to understand data warehouse steps.
    I studied some oracle documents and now I'm familiar with data warehouse
    but I have some question I couldn't find any answer in documents clearly.
    maybe my questions were so funny but if I can understand them I can go to next step in my job
    all questions are about OWB (embeded in oracle database 11g) (I ask this questions in Data Warehouse Forum but no body answer to these)
    *1- if I use this version of OWB, I encounter problem?*
    *2- it is better to use embeded version or stand alone version of OWB?*
    *3- in my schema in target used as data warehouse I should create objects manualy or OWB must create object in DB??*
    *4- if I create a map in target module how can I use tables in source module to create mapping for data flow??*
    *5- it's funny but when i create a map, I can't go to map editor to add objects to map?*
    BEST REGARD

    Pl do not post duplicates - Data Warehouse - your original post is in the correct forum for this topic.

  • CF 10 - IIS 7 - Set up an Access Data Source

    How do we set up an Access data source with CF10 and IIS 7 in a 64 bit environment.  
    I have been away from CF for several years and things are a little different now.
    I keep getting     
    Unable to update the NT registry.  Variable DRIVERPATH is undefined.
    -thx

    ColdFusion doesn't support MSAccess in 64bit. Additionally, Windows doesn't currently ship with a MSAccess .accdb driver.
    Please refer to http://blogs.coldfusion.com/post.cfm/coldfusion-10-64bit-and-msaccess

Maybe you are looking for