Oracle Data Warehouse DB and OLAP_OBIEE DB [OBIEE11g]

Hi Experts,
I have a problem regarding joining the 2 DB in an Analysis report.
The scenario is this:
Oracle Data Warehouse is DB_1 and OLAP_OBIEE is DB_2
DB_1
-col1
-col2
DB_2
-col3
In Analysis report when they are queried or run separately, obviously there will be a report generated.
But when this 2 DB columns are joined together in one Analysis report, I always get 'No Results'.
Please help how do I fix this issue :(
Thanks,

Hi,
check the following links:
http://www.rittmanmead.com/2007/10/reporting-against-multiple-datasources-in-obiee/
http://108obiee.blogspot.com/2009/01/fragmentation-in-obiee.html
http://gerardnico.com/wiki/dat/obiee/multiple_subject_area

Similar Messages

  • Tables between OBAW and Oracle Data Warehouse in OBIA rpd are different

    Hi,
    The tables in the Data Warehouse DB is different as compared to the tables in the physical layer of the Oracle data Warehouse in OracleBIAnalytiscApps.rpd.
    When I click update Row Count, it gives me below error message:
    There was an error while updating row count for "Oracle Data Warehouse"."Catalog"."DBO"."W_CTRY_REGN_D":
    [nQSError :17001] Oracle Error Code: 942, message: ORA-00942:Table or view does not exist at OCICallSTmTEXECUTE
    Whereas when I click on some other tables for Row count there is not error.
    Also, When I am trying to look for some specific table which exist in Data Warehouse (that has the data), it does not exist in the rpd.
    Kindly assit me.

    hi,
    Have a look
    Not able toView Data in Answers
    Let me know
    thanks,
    saichand.v

  • Designer Vs. Oracle Data warehouse builder

    Dear all,
    Currently I'm responsible of building a Data warehousing project using Oracle database. I'm trying to decide on a tool for modelling my datawarehouse. I have two options:
    1) Designer: we have some experience with this tool and we are using it for our main OLTP application.
    2) Oracle Data Warehouse builder: we are using this to design our ETL processes.
    I want to get some advice on whether the OWB is capable of modelling my datawarehouse and of doing a retrofit action. also, I try to standardize on the tools that are using in the Data Warehouse department (currently we are using only OWB).
    I will appreciate for any other advice to help in my selection process.
    Best Regards,
    Bilal

    Hi,
    In my experience this choice depends on the implementation of the datawarehouse. If you are building a "pure" Kimball style dimensional data warehouse you should be able to do this using OWB. I have architected such a DW in the past using only OWB, so I am speaking from experience.
    If on the other hand you are planning to implement an Inmon style CIF, if your requirements includes an operational data store (ODS), or if you for any other reason anticipate that you are going to be doing a lot of ER modeling, then I would not recommend using the current release of OWB for modelling. (Note however that there are significant improvements to the modelling capabilities in the Paris release of OWB, so this may change in the future)
    The advantage of improved maintainability when using a single tools needs to be weighted against the improved functionality if you choose a combination of the two. In the "two tool" scenario strict development and deployment routines need to be enforced to avoid that the model in Designer comes out of sync with the metadata in OWB. (Consider the effect of a developer making a change to a table definition in OWB and deploying it directly to the database without updating the model in Designer.)
    Hope this helps.
    Regards,
    Roald

  • PAS - Connection to Oracle data warehouse

    Hi,
    We are looking to connect to an Oracle data warehouse and have installed an Oracle client on the server.
    We created the link ID and are using the built in connection to Oracle OCI.DLL. However we get an error back saying that it could not load the OCI.DLL
    Any ideas?
    Colin Cooper

    Hi Robert,
    I'm confused Pedro. What is the NW Database Instance? Is this another installation of CE? Or is this the relational database running behind CE? More confusing is the fact that you say SQL Server is installed on the "Database Server". If SQL Server is on this machine, why is Oracle in the picture?
    Because of the system landscape this customer have, we use a distributed system instalation of NetWeaver (NW), that is composed by:
    - Central services instance (SCS)
    - Database instance (DB)
    - Primary application server instance
    The database instance is the one installed in the Database Server.
    Also MS SQL Server is installed in the same server, because it's used as sytem database for SSM (store all the metadata of SSM).
    Finnaly, Oracle is the database used by the Datawarehouse, the place we should connect to get all the data to load into PAS database. This is a different server.
    I understand that. My question is, does Oracle have a fix for this issue available? Have you checked metalink to see whether or not a recent Oracle patch set, or the most recent Oracle client, already has a fix for this issue?
    I did not check that. I will check with the customer what is the version currently used and if exists any fix this issue.
    I have a machine running Windows 2003 Server 64 with the 32bit Oracle client installed and have no issues with setting up an ODBC connection to Oracle. What I had to do though was ensured that I ran "C:\WINDOWS\SysWOW64\odbcad32.exe" and not the standard "%SystemRoot%\system32\odbcad32.exe" which is what the ODBC applet in control panel usually calls. Because you're running 32bit Oracle, you have to run the 32bit ODBC applet in order to successfully create the System DSN which can be called by PAS.
    That's exactly what I need. As soon I try this I give you a feedback.
    Thanks
    BR
    Pedro

  • Tools required to install to work with Oracle Data warehouse

    Could anyone please guide me...as I am quite new to Oracle...
    till now I was working on Data warehousing using Microsoft tools i.e. (SSIS, SSRS, SSAS, SQL Server)
    So Please tell me which are the tools required to install to work with Oracle Data warehouse
    I came to know about OWB and OBIEE, but not sure what exactly these are ?

    In order to create and develop a webcenter portal application, you only need to install JDeveloper and download the WeBCenter extensions. That's true.
    But this install does not include the WebCenter services like dicussions, document services, analytics,... If you want to use these and integrate them in your portal, then you would need to install the suite as described on my blog.

  • Data warehouse backups and read only tablespaces

    Hi all,
    I am working on a data warehouse database with following specs:
    Version: Oracle 10.2.0.3 Enterprise
    OS: Solaris
    App: Data warehouse
    We use RMAN to take 'level 0' & 'level 1' backups. We have block change tracking enabled and RMAN backups up data files and archive logs straight to tape.
    I am exploring ways of reducing the 'level 0' backups and was specifically focussing on using read-only tablespaces for this purpose.
    I have often seen it mentioned that a best practice in D/Ws is to store the old static partitions of fact tables in read only tablespaces so as to reduce the backup size.
    In case you have already implemented such a scheme, I would like to know how you have implemented it.
    I am thinking of the following mechanism:
    -- Start using backups at tablespace level rather than 'level 0' at database level.
    -- Record the latest SCNs of all datafiles prior to back up.
    -- If the latest SCN has not changed since last backup and the tablespace is in read only mode then
    -- Check if a backup copy of the tablespace has been done within the recovery window and is accessible.
    -- If the copy exists then don't backup that tablespace, else backup the tablespace.
    -- If the tablespace is read/write then back it up.
    I haven't delved into the low level details, but this seems to be lot of work. So just wanted to know from you if there's any ready made feature which makes all this easier.
    Many thanks in advance.

    Thank you so much for your help.
    Backup optimization was indeed the thing I was looking for. To be honest I had done a bit of RTFM, but I didn't check the advanced user guide.
    Although my specific question has been answered, it would be interesting to know what other things other people are implementing to reduce backups etc.
    I am also thinking of following options:
    -- Turn on index monitoring to get rid of unused indexes.
    -- Stop the backups of 'index' tablespaces.
    -- Archive off old data.
    Any other ideas for reducing DW backup size?
    Many thanks.

  • Efficiency of data warehouse sql and star/snowflake schema

    Hi,
    We are using 11.2.0.3 and need to improve query performance of reports.  data warehouse star/snowflake schema
    In addition to indexing, partitioning having star_transformation enabled etc I am condisriing impact of the following on query performance.
    central fact (over a billion rows) joins to a dimesnion customer ( few hundred thousand rows) which in turn joined to latest version of the dimesnion ( whichhas circa 30,000 rows).
    The table with few hundred thousand rows (customer dimesnion) must alwsys be queried as data stored aganist the version of customer applicable at the time - we just query latest_customer as users want to see
    latest version of customer attributes to stop data being fragemented across several rows in the report.
    Considering if would be more efficient to create a dimenson which is the equivalent of customer but also stores the latest version of the customer attributes on the on row - this would mean customer dimensuion would have far more columns but queries would could avoid additional lookup of this 30k row table.
    Thoughts are - would this be a material benefit?
    At monent users would query latest_customer to say get all customers belonging to a certain multiple chain.
    If change as above then they would query directly the customer dimension with few hundred thousand rows.
    Thoughts?
    Thanks

    We are using 11.2.0.3 and need to improve query performance of reports.  data warehouse star/snowflake schema
    That is NOT a realistic or even meaningful goal.
    And until you identify and document an actual PROBLEM or specific goal you should not even be considering possible solutions.
    Anything you do to improve one report might degrade the performance of several other reports.
    You need to start over and gather information about WHAT Oracle is doing for the reports now, HOW that work is being done and capture metrics that validate how the reports are currently performing.
    Your first step should be to document the performance you are getting now for each report.
    The second step would be to identify which of those reports is a possible target for tuning.
    The third step is to prioritize the reports: which is most important to tune, which is next, etc.
    Then you need to generate the execution plans for those reports to identify EXACTLY how Oracle is executing the queries now.
    At this point you should have enough information to know what your possible options are.
    So then you create a prioritized list of options. The top of the list should be additions to what you already have.
    1. New indexes - regular or bitmapped (if appropriate)
    2. Dropping indexes that aren't being used.
    3. Report-ready summary tables or Materializeds views.
    IMHO modifying your basic architecture should be your LAST resort and undertaken only if you can't solve your (unstated) problem using solutions that have less impact and risk.

  • How create oracle data Warehouse

    hi master
    sir i am use oracle 9i how i creat data Warehouse
    and how use oracle portal for net
    please give me idea how i start and which woftware i use
    thank's
    aamir

    Do you want to craete a Data Warehouse database or you want to learn on how to design a Data Warehouse ?
    Creating a Data Warehouse database is no different than creating an OLTP database, the only thing in mind should be is that DW are mostly read-only, so you should set initialization parameters suiting DW .

  • Difference between Data warehouse BO and SAP BO

    Dear All,
    can anyone let me know the difference b/n data warehosue Business Objects and SAP Business Objects.....and aslo can u suggest which is the better future among these two tools....
    explaining in detailed view is highly appreciable!
    Thanks & Regards,
    V.Ravi Teja

    Hello,
    in the SAP BusinessObjects portfolio there is no solution for a DWH. A DWH is a DataStore based on a Database technology like Oracle. SAP BusinessObjects doesnt have its own Database. SAP BusinessObjects Tools support loads of Databases like Oracle, DB2, MS SQL etc.
    What the SAP BusinessObjects portfolio has is an ETL tool to create a DWH - called "DataServices". Also there is a tool for creating a virtual DWH - called "Data Federator"
    SAP has its own Datastore for a DWH - called "SAP Business Warehouse" (SAP BW)
    Hope this helps.
    Regards
    -Seb.

  • Recommended throughput for Oracle data warehouse

    Hi, I know up front this is going to be a vague question...but I'm trying to determine approximate I/O bandwidth for a data mart server. Right now we're hosting 3 or 4 different marts on it, but that number is going to increase.
    Oracle's DW "2 day" class recommends starting with either maximum throughput from user queries, or basing it off of batch windows. Right now the server is barely used for end user queries, as we haven't yet implemented a BI tool to allow users easy access (that's underway right now). So I find it hard to base any info on that. However, it's on the way, and I'm in charge of the BI took (OBIEE). I'm having nightmares that we get OBIEE deployed, and our queries end up taking 5 minutes each to get answers... Right now, on the system basically by myself, if I do a simple "select sum(amount) from fact_ledger", where fact_ledger is a 1 Gb table (with 40 million rows), it takes almost a full minute to run. It feels like I could add this up by hand and get an answer faster...and this certainly doesn't compare with other Oracle marts / DWs I've worked on in the past.
    From a batch window standpoint, all I can say is that it feels really, REALLY too slow to me. Right now, some jobs that start with a 40 million row table and join it to 6 or 7 other small tables (looking up surrogate keys) and writing to a non-logged, non-indexed output table takes over 2 1/2 hours to complete. To me this should be a 15 minute job.
    We've asked IT to do a "root cause analysis" of why performance is so bad - but as part of that, the architecture group wants something more concrete than "it just feels way too slow". So does anyone have some general guidelines they can provide? I guess our detailed info would be:
    - three marts, each of which has a fact table around the 30 - 60 million row level
    - simple "join 30 million row staging table to look up surrogate keys" and writing results is taking 2.5+ hours
    - we expect at some point to have mabe 50 - 100 users running data concurrently (spread across the marts)
    - users will be performance both canned and ad-hoc analysis against it...and they are high level business users, aren't going to be happy with waiting 2 minutes for a simple answer
    My start was to swag this as requiring 6 CPUs or so, which would indicate (according to Oracle's best practice docs) of needing somewhere betweeen 1.2 GB/s to 2.4 GB/s throughput. I'm assuming if it takes almost a full minute to read a 1 GB table, that our IO is currently 60 to 120 times too slow. Does that make sense?
    Thanks and sorry for the lack of details...we just don't know yet.
    Thx,
    Scott

    Why don't you start by taking an AWR report from those two hours so you can see what is the bottleneck for your system ?

  • Oracle Data Guard: Physical and Logical

    I have a Primary database and have created a Physical Standby on another node. The physical standby is kept in synch via REDO Aply - online redo logs.
    QUESTION: is it possible to create a Logical Standby off of the Physical Standby? I dont think so since the logical is kept in synch from a primary via SQL- Aply. CAN SOMEONE PLEASE CONFIRM.
    I thought that a logical standby MUST be created from a Primary and not a Physical.
    Thanks!!

    Documentation is your friend,Orace does not not hide the information,how to create a logical standby:
    http://download.oracle.com/docs/cd/E11882_01/server.112/e10700/create_ls.htm#g105412
    Werner

  • Oracle Data Access Components and Windows XP Professional x64

    To start I'm running Windows XP Professional x64, using IIS 6, and Visual Studio 2005. I have a web service project that uses ODAC 11g, which runs just fine when I run it in debugging mode. However when I publish the project and try to access it through IE, I get an application error saying that Oracle.Web or one of its dependencies could not be loaded. Now I suspect that this might be happening becuase I'm running on a 64-bit machine, but I'm not convinced. The web service worked just fine on a 32 bit windows running IIS 5. If anyone could shed some light on this that would be awesome. Thanks!

    hi ,
    i have more than one question indeed.
    1.i installed personal oracle v 8.1.7 for windows nt/2000 on
    windows xp profissional platform and it works great.When i
    tryed to install developer 2000 6i a microsoft runtime error
    is produced. is thier any solutions for such problem?
    2.when i tryed to install personal oracle on another machine with the following specs:
    -p4 1.4 ghz intell proccessor
    -gigabyte motherboard built-in sound
    -128 mb ram
    -S3 savage4 32mb vga card
    the oracle universal installer never runs.

  • Oracle Development Survey: Data Warehouses Customers

    At the start of most data warehouse projects, or even during a project, I am sure you as customers try to find answers to the following questions to help you plan and manage your environments:
    * Where can I find trend and comparison information to help me plan for future growth of my data warehouse?
    * How many cpu's do other customers use per terabyte?
    * How many partitions are typically used in large tables? How many indexes?
    * How much should I allocate for memory for buffer cache?
    * How does my warehouse compare to others of similar and larger scale?
    The data warehouse development team, here at Oracle would like to help provide answers to these questions. However, to do this we need your help. If you have an existing data warehouse environment, we would like to obtain more technical information about your environment(s) by running a simple measurement script and returning the output files to us, here at Oracle. This will allow our developers to provide comprehensive documents that explain best practices and get a better understanding of which features our customers use the most. This will also allow you as Customers, to benchmark your environments compared to other customers’ environments.
    From a Company perspective we are also interested to get feedback on features we have added to the database, are these features used, how are they used etc. For example we are keen to understand:
    * Which initialization parameters are most frequently used at what values?
    * How many Oracle data warehouses run on RAC? on single nodes?
    * Is there a trend one-way or the other, especially as data volumes increase?
    * Does this change with newer releases of the database?
    All results from these scripts will be held confidential. No customers will be mentioned by name; only summaries and trends will be reported (e.g., “X percent of tables are partitioned and Y percent are indexed in data warehouses that are Z terabytes and larger in size.” or “X percent of Oracle9i and Y percent of Oracle10g data warehouses surveyed run RAC”). Results will be written up as a summarized report. Every participating customer will receive a copy of the report.
    Terabyte and larger DW are the primary interest, but information on any data warehouse environment is useful. We would like to have as many customers as possible submit results, ideally by the end of this week. However, this will be an on going process so regular feedback after this week is extremely useful.
    To help our developers and product management team please download and run the DW measurement script kit from OTN which is available from the following link:
    http://www.oracle.com/technology/products/bi/db/10g/dw_survey_0206.html
    Please return the script outputs using the link shown on the above web page, see the FAQ section, or alternatively mail them directly to me: [email protected].
    Thank you and we look forward to your responses.
    Message was edited by:
    klaker

    At the start of most data warehouse projects, or even during a project, I am sure you as customers try to find answers to the following questions to help you plan and manage your environments:
    * Where can I find trend and comparison information to help me plan for future growth of my data warehouse?
    * How many cpu's do other customers use per terabyte?
    * How many partitions are typically used in large tables? How many indexes?
    * How much should I allocate for memory for buffer cache?
    * How does my warehouse compare to others of similar and larger scale?
    The data warehouse development team, here at Oracle would like to help provide answers to these questions. However, to do this we need your help. If you have an existing data warehouse environment, we would like to obtain more technical information about your environment(s) by running a simple measurement script and returning the output files to us, here at Oracle. This will allow our developers to provide comprehensive documents that explain best practices and get a better understanding of which features our customers use the most. This will also allow you as Customers, to benchmark your environments compared to other customers’ environments.
    From a Company perspective we are also interested to get feedback on features we have added to the database, are these features used, how are they used etc. For example we are keen to understand:
    * Which initialization parameters are most frequently used at what values?
    * How many Oracle data warehouses run on RAC? on single nodes?
    * Is there a trend one-way or the other, especially as data volumes increase?
    * Does this change with newer releases of the database?
    All results from these scripts will be held confidential. No customers will be mentioned by name; only summaries and trends will be reported (e.g., “X percent of tables are partitioned and Y percent are indexed in data warehouses that are Z terabytes and larger in size.” or “X percent of Oracle9i and Y percent of Oracle10g data warehouses surveyed run RAC”). Results will be written up as a summarized report. Every participating customer will receive a copy of the report.
    Terabyte and larger DW are the primary interest, but information on any data warehouse environment is useful. We would like to have as many customers as possible submit results, ideally by the end of this week. However, this will be an on going process so regular feedback after this week is extremely useful.
    To help our developers and product management team please download and run the DW measurement script kit from OTN which is available from the following link:
    http://www.oracle.com/technology/products/bi/db/10g/dw_survey_0206.html
    Please return the script outputs using the link shown on the above web page, see the FAQ section, or alternatively mail them directly to me: [email protected].
    Thank you and we look forward to your responses.
    Message was edited by:
    klaker

  • Oracle Development Survey on Data Warehouses: How Does Yours Compare?

    At the start of most data warehouse projects, or even during a project, I am sure you as customers try to find answers to the following questions to help you plan and manage your environments:
    * Where can I find trend and comparison information to help me plan for future growth of my data warehouse?
    * How many cpu's do other customers use per terabyte?
    * How many partitions are typically used in large tables? How many indexes?
    * How much should I allocate for memory for buffer cache?
    * How does my warehouse compare to others of similar and larger scale?
    The data warehouse development team, here at Oracle, would like to help provide answers to these questions. However, to do this we need your help. If you have an existing data warehouse environment, we would like to obtain more technical information about your environment(s) by running a simple measurement script and returning the output files to us, here at Oracle. This will allow our developers to provide comprehensive documents that explain best practices and get a better understanding of which features our customers use the most. This will also allow you as Customers, to benchmark your environments compared to other customers’ environments.
    From a Company perspective we are also interested to get feedback on features we have added to the database, are these features used, how are they used etc. For example we are keen to understand:
    * Which initialization parameters are most frequently used at what values?
    * How many Oracle data warehouses run on RAC? on single nodes?
    * Is there a trend one-way or the other, especially as data volumes increase?
    * Does this change with newer releases of the database?
    All results from these scripts will be held confidential. No customers will be mentioned by name; only summaries and trends will be reported (e.g., “X percent of tables are partitioned and Y percent are indexed in data warehouses that are Z terabytes and larger in size.” or “X percent of Oracle9i and Y percent of Oracle10g data warehouses surveyed run RAC”). Results will be written up as a summarized report. Every participating customer will receive a copy of the report.
    Terabyte and larger DW are the primary interest, but information on any data warehouse environment is useful. We would like to have as many customers as possible submit results, ideally by the end of this week. However, this will be an on going process so regular feedback after this week is extremely useful.
    To help our developers and product management team please download and run the DW measurement script kit from OTN which is available from the following link:
    http://www.oracle.com/technology/products/bi/db/10g/dw_survey_0206.html
    Please return the script outputs using the link shown on the above web page, see the FAQ section, or alternatively mail them directly to me: [email protected].

    969224 wrote:
    Hi Guys, just a quick question. when we have a primary key on 4 coloumns and we have, say 20 million rows and we want to add one extra row. How does oracle check whether the data on the primary key is unique to the record being added compared to the 20 million rows. Does it actually compare the record being added to all the rows present in the table?
    Edited by: 969224 on May 10, 2013 8:14 AMNot the whole row, it compares the 4 columns in the INDEX against the 4 columns in the new row.

  • Oracle based data warehouse creation

    Hi All ,
    I need to know Best Infrastructure for creating data warehouse. Our client is currenlty managing data in SQL Server,Oracle and XL sheets. Now they want to implement data warehouse,BI and Data mining.
    Scalability is our main concern because size of data will be in Tera Bytes (10 TB) in next few years. Scope is analysis,Impact analysis,Forecasting,GIS based analysis and mining(Trend analysis).
    Please suggess me on following point (I mean help me in tool selection):
    A. Storage
    B. Operating System, Database Server
    C. Other Server (Warehouse, Mining ,BI).
    D. Tools( Database ETL,OLAPReporting and Mining)
    I m working with Hyperion BI, I m new in data warehouse. This would be great if I get Cost of tools (for Perpectual licenses).
    Thanks in Advance
    Shiv

    Oracle Retail Data Model is independent of RDW and RGBU applications. It can be used with RMS and Point of Sale, but there is no pre-built ETL to load the data included with the product, although there are some options available thru partners.

Maybe you are looking for