Use Data Warehouse for EBS 12.1.1 Data

Hi,
I want to create a data warehouse for our EBS 12.1.1 system.
I read some article about it, but i don't understand Where should i start?
please tell me what should i do???
i install oracle warehouse builder but i couldn't used it.
Best Regard

Thanks Eric for the response.
Our requirement is, in a Production instance, not to allow a DB user (APPS/SYSTEM) to see the confidential data. For example, a user who has the APPS schema password should not be able to see the salary details. However allowing the application like forms/reports/jsp to present the actual data and not the masked data.
I understand from your reply that the data masking feature CANNOT be used for this purpose as it is rebiulding the tables with the masked data. However we understand that the same feature can be used in a TEST/DEV environment so that we can restrict the developer who is working on the TEST/DEV instance, from accessing the confidential data.
Any pointers to acheive our requirement will be highly helpful.
Thanks for your time.
Ramana.

Similar Messages

  • Build a data warehouse for crystal report use

    OS: AIX 5.3 with Oracle 10g R2/Baan 5c
    I proposed my colleage to create a data warehouse for his heavily use of crystal report activity, so that his activites will not drag the normal database performance. The data will collect from the current Oracle 10 g database. Because I have never physically worked on the data warehouse, where should I start to look for information to proceed? I have an extra Sun Solris 8 box with a spare HD. I think it will be better to build the warehouse on that server, am I right? or the 10g R2 provides the capability to host another warehouse activity -- create a materialized view within? Thanks,

    Thanks for the help Ted
    Works like a charm now.
    Or atleast my very very simple XML report works.
    Now I just need to play around and see what I can do and what I cant.
    I just want to make sure I understand how this works.
    1) I start of with a XML file (or URL or other data source as explained in your BLOG)
    2) I then have to create blank report which will have its data source set to my XML file.
        This report (rpt file) will always have its data source set to the original XML location
    Can this be dynamically changed, or does a new rpt file need to be created each time
        (or atleast I will have to re-run "CreateXMLInlineSchemaReport" from your BLOG)
    3) I can now use this rpt file as I would any other rpt file but its data source will be the original XML file
       - If I update or change the XML file, the report will dynamically reflect these changes.
    4) To change the data source, I need to recreate a new rpt file as per point 2. Which will now be blank.
    5) Filtering.... if I want to filter the XML data do I use normal parameters. If I want to filter the data before
        I pull it do I need to change anything. In otherwords if I have a XML file with 1000 records, but I only want say
    5 of them, how do I only pull those 5 without first pulling the whole 1000 records. Hope that makes sense.
    As I said I am very new to this so some things still fly over my head, but I do play around abit so hopefully
    will find what I am looking for.
    Cheers
    Darren
    PS any other links to info on creating and running reports of XML data sources would be greatly appriciated

  • Recommendation for setting up Data Guard for EBS R12 12.1.3

    Hello Experts,
    I would like to get some experts opinion and recommendation on setting up Data GUARD for EBS R12.1.3 running on Linux based systems.
    Firstly I would wanna let you all know that I am just geek in this subject matter so do excuse me for any mistakes.
    While we are planning on setting up Data Gurad fro R12 what are the best practices to consider ?.
    We have 2 node rac enabled database and would wanna have non-rac based physical standby, based on this I believe that having db tier on remote or standby site is enough ?
    if db node is enough for standby location what are things to be done when a fail over / switch over / role transitioned to standby for the app node from primary site to point db on the
    standby location or do we have to have both apps and db tier @ stand by location
    --However having a apps tier/node at standby will be of no use as we are planing to run db on mounstate meaning not a active data guard setup.
    Please do reply with your recommendation/suggestions/pointers
    Thanks in Advance.

    Service Provider Access resulted in exception 'oracle.apps.fnd.soa.util.SOAException: SystemError: Error while sending message to server. http://ec2-107-22-78-224.compute-1.amazonaws.com:8000/webservices/SOAProvider/EbizAuth?Generate=1132&soa_ticket=Tu2z7GYoWPwq-VdgimKRFg..' when attempting to perform 'GENERATE'. Please view Service Provider logs for more details
    Can you find any details about the error in the log?
    I have looked at the following Note as well:
    Error using the Generate WSDL Button in Oracle E-Business Suite Integrated Soa Gateway Release 12.1.1 [ID 1090946.1]Did the doc help?
    Any one know what may be causing this issue? Do we have to do additional setup for SOA gateway after the standard install of the OVM? I tried to follow the steps in the note above but I do not see any entry for "<jdbc_url oa_var="s_apps_jdbc_connect_descriptor"/>" in the file data-sources.xml.
    Anyone have any ideas?Have you reviewed these docs?
    Oracle E-Business Suite Integrated SOA Gateway Troubleshooting Guide, Release 12 [ID 726414.1]
    Oracle E-Business Suite Integrated SOA Gateway 12.1.1 Consolidated One-Off [ID 815196.1]
    Thanks,
    Hussein

  • Where is Employee Nationality column in data warehouse for OBIAPPS 11.1.1.7.1 ?

    Hi,
    Where is Employee Nationality column in data warehouse for OBIAPPS 11.1.1.7.1 ?
    OS: windows 2008 R2
    BI APPS: 11.1.1.7.1
    ODI: 11.1.1.7.0
    Offering: HCM
    Thanks
    Rafeek Abd Elmonsef

    Hi,
    I checked BI APPS 7.9.6.4 Nationality column is available in W_EMPLOYEE_D table but in version OBI APPS 11.1.1.7.1 I did not able to find NATIONALITY Column in W_EMPLOYEE_D ?
    Thanks
    Rafeek

  • RE: data warehouse for apps

    Is there a way to prepare for a data warehouse for oracle Apps?
    Specifically is there a product that can cater to this?
    Also does anybody have any experience with this at their site?

    Check your Flat file structure (header) and the data source or infosource structure.. I think that the structures are not aligned here...

  • Data warehouse Loader did not write the data

    Hi,
    I need to know which products are the most searched, I know the tables responsible for storing this information
    and are ARF_QUERY ARF_QUESTION. I already have the Data Warehouse module loader running, if anyone knows
    why the data warehouse loader did not write the data in the database, I thank you.
    Thank.

    I have configured the DataWarehouse Loader and its components.Even I have enabled the logging mechanism.
    I can manually pass the log files into queue and then populate the data into Data Warehouse database through scheduling.
    The log file data is populated into this queue through JMS message processing" and should be automated.I am unable to
    configure this.
    Which method is responsible for adding the log file data into loader queue and how to automate this.

  • Do I need Management Data Warehouse for SAP on SQL?

    Hi experts,
      I am reading Jonathan Kehayias's wait statistics guide and I see he talk about Management Data Warehouse. Should I use Management Data Warehouse in SAP on SQL environment? I see many monitor objects are already implemented in DBACockpit.

    Hi Dennis,
    In case you have your SAP system on MSSQL, you can definitely use Management Data warehouse
    to streamline performance troubleshooting.
    SQL Server 2008 Management Data Warehouse
    Hope this helps.
    Regards,
    Deepak Kori

  • How to export data using application adapter for EBS?

    my scenario is : there're some external apps need to communicate with EBS,
    and the intermedia layer supposed to be implemented by SOA suite/Fusion adapters.
    my confusion is:
    in oracle integration repository, most of APIs are CUD (create, update, delete) without R (read). but many of our usecase was querying data from EBS.
    my question is : although, considering performance and unpridictable query object and granularity,it's not a good idea using app adatper to querying data from EBS via app api, but if I using db adapter to query data directly, it seems not easy as expected because it need to know the detail table structure and relationship and to build sql by hand.
    so, is there other proper way to get outbound data integration using fusion middleware? additional, I'm not clear about the ODI functionality , is it possible and easy to use ODI perform this task, eg, is ODI has enough API to interact with fusion adapter?

    In the Adapter Wizard look under "Other Interfaces/Custom Objects/PLSQL APIs". There are many many "Get" APIs here. None of these are identified or explained in the Integration Repository but the Trading Community Architecture - Technical Implementation Guide does give a little bit of information about what they do.

  • Can i use one hardrive for time machine and other data?

    i want to buy an external drive for my macbook pro retina. i want to knwo if i were to buy an external hardrive big enought ot back up my 768 on my retina could i also put other media on it (like video and movie) without it overwriting or something like that

    screenfreak wrote:
    i want to buy an external drive for my macbook pro retina. i want to knwo if i were to buy an external hardrive big enought ot back up my 768 on my retina could i also put other media on it (like video and movie) without it overwriting or something like that
    Buy one hard drive for TimeMachine, large as possible for maximum amount of saved states, keep it connected more often so it updates more often to recover deleted files.
    Buy one hard drive the same size or slightly larger to clone your boot drive to it, it will be a hold option key bootable clone in case your boot drive fails to boot. Update it before doing anything major to your machine. Keep it disconnected in case of malware so it doesn't get infected or delete your files.
    Buy one hard drive for extra storage if your boot drive is getting full for instance.
    Buy one hard drive to backup the extra storage drive and treat it like the clone, make sure you have clean extra storage drive before connecting and updating it.
    Label and date each drive, your memory will fail you and you'll accidentally overwrite a drive.
    Keep your drives separate and in their respective purposes, in other words do not use partitions or use TimeMachine to backup other drives.
    Never use Filevault, it's cracked, it's worthless, impossible to recover data and you'll have to give up the password for repair or customs/searches anyway. Use a external hardware based encrypted drives and USB Iron Keys. This way your sensitive data is off the machine and can be used with another machine.
    Despite how many drives and backups, it's good to also make burned and labeled DVD's of your most important files like music for instance when you first get them. It's because sectors fail on hard drives and takes out your data and thus on all backups made afterwards. With burned DVD's it's permanent, unless you scratch it, melt it or loose it.
    Hardware also fails, more often than software because it's moved/abused while in operation.
    If you lose your TM drive, you ALSO lose what's on the other partitions on the drive or data that TM drive was backing up from other drives.
    Drives are cheap, data is not. Protect the data. Because it's a fortune trying to recover lost data off a dead drive.

  • IS IT WORTH TO USE IOS 8 FOR IPHONE 4S AS ON DATE ?

    Kindly guide whether ios 8 is to use now for iphone 4s or hav to wait for some improvements for 4s in near future ? as I came to understand iphone 4s users facing some difficulties. pls guide.
    thanx n regds.

    Even to get listed in the System Share Menu?
    I used the code below in an Air for Android project to do that:
      <intent-filter>
       <action android:name="android.intent.action.SEND" />
       <category android:name="android.intent.category.DEFAULT" />
       <data android:mimeType="text/plain" />
       </intent-filter>
    I was hoping for something similiar for iOS 8.

  • Using Frame API for C, Need to change Data size for a TX Object

    I need to Transmit different messages to the same ArbitrationID address. The Transmited messages have variable lengths. Can I call ncConfig on an object to change the NC_ATTR_CAN_DATA_LENGTH even after the call to ncOpenObject. Or do I have to Close the object and reopen it after the call to ncConfig?
    Thanks!

    Hello Genesis,
    your second assumption is right. You do have to close, reconfigure and reopen the object in order to change the data length.
    As long as you're using the CAN Object in the 'Transmit Data by Call' mode only, you might want to consider using the Network Interface for transmitting. When you use the Network Interface, you can change the ArbitrationID for each write call.
    -B2k

  • Using Cursor and FOR LOOP to INSERT the data into table

    Hi all,
    I have SELECT statement that returns 3 rows:
    PROCESSNAME
    PROTDATE
    IMM
    2013-12-18
    Metrology
    2013-11-18
    CT
    2013-12-04
    SELECT  processName, MAX(NVL(protStartDate, protCreateDate)) AS protDate
        FROM TABLE(SEM_MATCH("{
                ?ipc rdf:type s:Protocol .
                ?ipc s:protocolNumber ?protNum .
                ?ipc s:protocolCreateDate ?protCreateDate .
                OPTIONAL {?ipc s:protocolSchedStartDate ?protStartDate }
                ?ipra rdf:type s:ProcessAggregate .
                ?ipra s:hasProtocol ?iprot .
                ?iprot s:protocolNumber ?protNum .
                ?ipra s:processAggregateProcess ?processName.
        }",sem_models("PROTS", "LINEARS"),NULL, SEM_ALIASES(SEM_ALIAS("","http://VISION/Data/SEMANTIC#"),SEM_ALIAS("s","http://VISION/DataSource/SEMANTIC#")),NULL))
            Group by processName
    Now I need to INSERT these values into the table along with the other values.
    these other values come from different table.
           INSERT INTO MODEL_CLASS_COUNTS (MODEL_NAME, CLASS_NAME, INS_COUNT, COUNT_DATETIME, PROCESS_NAME, PROT_DATE)
           VALUES
           ("$MODEL",     
                "${i}",
            (SELECT COUNT (DISTINCT S)  FROM TABLE(SEM_MATCH(
                            "{?s rdf:type :${i} . }",SEM_Models("$MODEL"),NULL, SEM_ALIASES(SEM_ALIAS("","http://VISION/DataSource/SEMANTIC#")),NULL))),
             SYSTIMESTAMP, %%here need to insert PROCESSNAME, PROTDATE%%
    t was giving me error:
    PL/SQL: ORA-22905: cannot access rows from a non-nested table item
    so i enclosed sparql query into single quotes.
    The code is as follows:
    declare
    type c_type is REF CURSOR;
    cur c_type;
    v_process varchar2(200);
    v_pdate varchar2(200);
    begin
    open cur for
           ' SELECT processName,  MAX(NVL(protStartDate, protCreateDate)) AS protDate   <-- it's complaining about this being too long identifier, i think...
            FROM TABLE
              (SEM_MATCH (
                            ?ipc rdf:type s:Protocol .
                            ?ipc s:protocolNumber ?protNum .
                            ?ipc s:protocolCreateDate ?protCreateDate .
                            OPTIONAL {?ipc s:protocolSchedStartDate ?protStartDate }
                            ?ipra rdf:type s:ProcessAggregate .
                            ?ipra s:hasProtocol ?iprot .
                            ?iprot s:protocolNumber ?protNum .
                            ?ipra s:processAggregateProcess ?processName.
                        }",SEM_Models("XCOMPASS", "XPROCESS"),NULL,    
              SEM_ALIASES(SEM_ALIAS("","http://VISION/Data/SEMANTIC#"),
              SEM_ALIAS("s", "http://VISION/DataSource/SEMANTIC#")),NULL))
               Group by processName';  
    loop
    fetch cur into v_process, v_pdate;
    exit when cur%NOTFOUND;
    --here I need to insert v_process , v_pdate into my table along with other values...
    dbms_output.put_line('values for process and prod_date are: ' || v_process || v_pdate );
    end loop;
    close cur;
    end;
    exit;
    Now, I get an error:
    ORA-00972: identifier is too long
    Does anyone know way around this?

    Hi,
      I tested something similar with insert into select  and it worked fine :
    insert into t_countries(ID,CITY ,POPULATION, DESCRIPTION, located, insdate )
    SELECT 1 id, city, o , city||' is a nice city' description,  max(nvl(locatedAt,'unknown')) as located,
      SYSTIMESTAMP
      FROM TABLE(SEM_MATCH(
        '{GRAPH :gCH {<http://www.semwebtech.org/mondial/10/countries/CH/> :hasCity ?cityID .
           ?cityID :name ?city .
           OPTIONAL{?cityID :locatedAt ?locatedAt .}
           ?cityID :population ?o .
        SEM_Models('VIRT_MODEL_MONDIAL'),
        SEM_Rulebases(null),
        SEM_ALIASES(SEM_ALIAS('','http://www.semwebtech.org/mondial/10/meta#'),
        SEM_ALIAS('prv','http://www.semwebtech.org/mondial/10/countries/CH/provinces/')
        null))
        group by city,o
        order by city;
    Or with execute immediate :
    declare
      v_country varchar2(200) :='http://www.semwebtech.org/mondial/10/countries/F/';
      v_text varchar2(2000);
    begin
    v_text := 'insert into t_countries(ID,CITY ,POPULATION, DESCRIPTION, located, insdate )
    SELECT 1 id, city, o , city||'' is a nice city'' description,  max(nvl(locatedAt,''unknown'')) as located,
      SYSTIMESTAMP
      FROM TABLE(SEM_MATCH(
        ''{<'||v_country||'> :hasCity ?cityID .
           ?cityID :name ?city .
           OPTIONAL{?cityID :locatedAt ?locatedAt .}
           ?cityID :population ?o .
        SEM_Models(''VIRT_MODEL_MONDIAL''),
        SEM_Rulebases(null),
        SEM_ALIASES(SEM_ALIAS('''',''http://www.semwebtech.org/mondial/10/meta#'') ),
        null))
        group by city,o
        order by city';
        dbms_output.put_line(v_text);
        delete from t_countries;
        execute immediate v_text ;
        commit;
    end;
    Marc

  • Use of warehouse for 'as of' balance custom fields

    we captured custom fields on each account/contact for current balance, current deposits, current trade volume
    we are updating these fields daily based on the transactional system integration
    how/will i be able to report on the day over day #'s or month over month given the built in warehouse -
    client wants a report to track the historical progression of these balances. Do we need to capture any additional fields?
    I thought this was a better approach then capturing the individual transactions (e.g. deposits, withdrawals, trades) as
    custom objects. Please would someone confirm?
    Edited by: user8903936 on Dec 8, 2010 11:01 AM

    Hi,
    my Ora-Trigger-Solution works well on R/3 development system.
    On quality assurance system it dumps with ORA-04098.
    It looks like, that required object privileges have been never granted.
    Do someone know, what priviliges or authorisations have to be maintained and where to maintain this?
    Thanks and regards,
    Wolfgang

  • Can SAP BI (BW) stand alone as a data warehouse for Non-SAP Source Systems.

    I recognize that the answer to this question depends on the version of SAP BW we are discussing. My question is one being posed to determine how SAP BW fits in with the BI world and the other tools available.
    - Can a company implement SAP BI without also having the SAP ERS applications?
    - If so, will SAP BI effectively support 3rd party source systems and what is needed to make it effective (NetWeaver XI, NetWeaver 2004s)?
    Thanks for your expertise.
    Denny

    hi dear,
    clearly the most valuable input given by BW is a native integration with all SAP world...
    well, you can manage a BW without having a SAP ERP as source system:
    http://help.sap.com/saphelp_bw33/helpdata/en/80/1a618ae07211d2acb80000e829fbfe/frameset.htm
    (DBconnect, BAPI...)...
    Hope it helps!
    Bye,
    Roberto

  • Data collection for cost object hierarchy - no data at order level!

    Hi everyone,
    When using transaction KKP6 to see my hierarchy for cost object, the system displays all the values that are allocated on my hierarchy nodes, but I do not have the values that should come from the production orders (repetitive manufacturing with production version).
    I use transaction KKRZ to collect data for cost object hierarchy, but it seems that the system does not retrieve any information from the orders.
    Would someone know about this point customizing steps to follow to retrieve my orders values?
    Thanks in advance for your answer. Rewards guaranteed!
    Ronan

    Hi Amadeus,
    To do so, I have used transaction OKZ5 in the customizing and ticked "always cost object hierarchy" and "distribution".
    By doing so, you enable the system to distribute costs within a cost hierarchy. Otherwise the system will only determine costs on the hierarchy.
    Try this, it may work for you as well. During your period-end closing, you sould now use "cost object hierarchy" instead of product cost collector.
    Regards,
    Ronan

Maybe you are looking for