Avoid EBS Vision Database data on Full Load in OBIEE

I've run Full Load after stressful installation and integration OBIEE 11g v.11.1.1.5 and OBI Applications.
The data loaded more or less OK, some parameters settings and application tuning need to be completed and it is not the issue.
The problem is that data from Vision database appear in Analytics Dashboards, specifically in GL Books drop-down LOVs.
I understand these values loaded from flat files because we don't have any tables from Vision.
We not loading Vision data and not need to show any information related to it.
My question is how avoid loading Vision data from flat files on the Full Load performing in DAC and Informatica.
Oleg.

Hi Sandeep,
You mean in Oracle eBS? Than it would be something like this;
select fur.user_name user_id
, fur.description user_name
, rsp.responsibility_name
, rsp.description responsibility_description
from apps.fnd_responsibility_vl rsp
, apps.fnd_user_resp_groups furg
, applsys.fnd_user fur
where rsp.responsibility_id = furg.responsibility_id
and furg.user_id = fur.user_id
and fur.user_name like nvl(:p_user_id,'%')
and rsp.responsibility_name like nvl(:p_resp_name,'%')
and furg.end_date is null
order by rsp.responsibility_name;
If you mean Oracle BI EE, check the repository via Manage, Security.
Hope this helps.
Good Luck,
Daan Bakboord
http://obibb.wordpress.com

Similar Messages

  • Embedded Database - data does not load

    Hello.
    I am currently using NetBeans 6.5 and am experimenting with the CarsApp tutorial.
    Whenever I create an application using the server database, the data loads automatically when the application starts.
    However, when I create the application using an embedded database, the data does not load and I just get a blank table.
    For more information, go to: http://forums.sun.com/thread.jspa?threadID=5366394&tstart=0
    Thank you in advance!

    I'm sorry, I didn't know it was so looked down upon.
    I thought that since this topic pertained to both Destop Applications and JDBC, I could get the best and quickest answer if I posted in both.
    Apparently, that was a mistake.
    Thanks, anyways!

  • Data Source - Full Load

    Hi Friends, Please help me with this requirement , I have a Standard content datasource that have plant, factory, Industry units and Machine Areas fields with  "valid to" and "valid from date". Now this datasource does not support the delta,
    1. can you help me how to make the full load on monthly basis, I have calendar day in DSO which i map to valid from field
    2. Is there any way to make this Standard content <b>datasource delta capable</b> to load delta to ods. No OSS notes available for this
    Please send me step bby step code to make the delta
    Thanks
    Poonam

    It is not possible to make generic datasource since too many tables are having the data fields. please suggest any other method
    Thanks
    Poonam

  • Problem with Data source- full load is not extracting where as delta can.

    Hi All,
    I have a problem in loading of the region values into BW.
    Data Source 0CRM_OPPT_I is used to fetch opportunities ( same as Sales Orders in ECC) from CRM system.Data source has a field 0CRM_TR which gives the values of region values pof the employee responsible. problem here is that CRM_TR value is not being fetched when the record was first created, but the value is being flown to BW in the delta run i.e. when a change happened in record.
    Please anyone help us out in resolving this issue.
    Regards,
    Haaris.

    Hi Hari.......
    Hav u done init........if yes........What u hav done....I mean.....init with or w/o data transfer.....?
    It has picked records or not? if yes...........after init.........hav u checked the Delta queue(RSA7) in the source system...........is the Total field is 0 or not.........if 0..then....
    Look if u r using Direct delta.then after init.any changes or new records will get recorded in the Delta queue(RSA7) directly..........but if the Update mode is Unserialized V3 update or Qued delta.then data will accumulated in the Update table and the Extraction Queue respectively...........in that case u hav to schedule V3 jobs to pull them in RSA7............otherwise delta load will pick 0 records only........
    Check the update mode of the Datasource in LBWE.......
    Check this..........it may help u........
    Steps to do init load from CRM
    Regards,
    Debjani.......

  • Vision database on fusion instanc

    Hello Experts,
       currently i have installed fusion application. Can anyone guide me whether i can find 'Vision Data base like R12' in fusion? I have checked with application as well as database with Apps access but I am not able to find any single row. Whether DBA missed out something while installation? Any patch is needed? any profile option?
    Kindly advise me whether such data is available ? If yes, where to find that?
    Thank you in advance.
    Regards,
    Basaw

    There is no "Vision database" available for Fusion. See from support system:
    Is There a Demo Database For Oracle Fusion Applications Similar to EBS Vision Database (Doc ID 1901090.1)
    Jani Rautiainen
    Fusion Applications Developer Relations                              
    https://blogs.oracle.com/fadevrel/

  • Init & Delta After Full Load

    hii Experts
    I have one doubt please clarify that.
    If i have a Cube and i m pulling data on basis of Full Load . Then i switch to delta load, before that i took the INIT load without data transfer.
    But after a couple of days i need to pull the full load again for that particular period, for that i delete the INIT request & also the Delta request from Cube & PSA. & after full load i want to pull the delta load.
    Now my doubt is how does this process impact on loading process?Is the previous init & delta also have some impact on the current delta loading???
    Thanks in Advance
    Neha

    Hi,
    There is absolutely no problem in doing a full loaf after init delta.
    doing full load doesnt disturb the delta status.
    Here i explain how an init and delta works
    Init and Delta :
    Once you are done with fetching all data from source using fullload, you have to run init.
    all the new records that appear in the source system after init will sit in the delta queue.
    when ever you run delta after the init, all the new records will coime to Bi and once this delta runs succesfully, these records will be cleared from the queue in source system and the new records from then comes to the queue.
    this way, we can be sure that there is no loss of data.
    Full load after delta :
    now comming to your answer, if you run a full load, the data is not fetched from the delta queue and it doesnt disturb the queue in any way. it only brings all the data from soure(according to selection condition). so the delta is not disturbed in any way.
    Repair Full  Request :
    The repair full request acts in the same way as full load in fetching data from source.
    the only difference is in updating the target.
    Rather than updating the data in the request, it checks out for the changes in already loaded data and update if there are any changes (Repairing the data already loaded.
    That means there is no impact of previous init without data transfer in the current delta loading?
    If you go thru the above explanation you can get this answer easily.
    Init is used to set the delta pointer due to which the present delta is working fine.
    Hope this answer helps you,
    Thanks and Regards,
    Srinath.

  • BI7 : Change full load to delta load - things to do?

    http://wiki.sdn.sap.com/wiki/display/profile/2007/04/30/Differencebetweenfull,intialanddeltaupdatemodes
    Difference: Full load, Delta Load & INIT load in BW
    Hi.
    Since am not familiar with BW request clarity as to how to achieve the following :
    We are on BI 7.
    On our BW server, One of the Existing process chain loads data as full load daily Till date.
    This schedule runs for about 4 to 6 hours during the nights. And the data seems to be growing.
    And doing full load, now seems, not that meaningful for data analysis.
    So we now feel we need not do a full load every day. But do only the incremental load.
    How to make this effective in our existing process chains.
    I went thro the 2 links above, but a few things am not clear.
    What exactly needs to be done to change a full load into a delta load.
    Also do not wish to have any data loss happening.
    We run the schedules on a particular process chain for BW around 1 am.
    Say from tomorrow I want to have the data loading as delta.
    What exactly are the steps to be followed.
    Precautions to be taken
    So that we have the correct data load and no data is lost - while doing this.
    Things to do am not clear :
    do initialization with data transfer / or without data tranfer ?
    The make it delta?
    So that from tomorrow all data loading becomes delta ?
    Or
    since in our scenario,
    daily basis it is full load.
    Just make it delta load,
    so that whatever changes will run in the next days schedule.
    What all things are to be done stepwise.
    if anyone could advise.
    That would be very helpful.
    Also any documentation which explains all these if someone could share the link please ?
    Many thanks
    indu
    Edited by: Indumathy Narayanan on Aug 26, 2011 5:50 PM

    Hi,
    Kindly look at following links as well
    http://help.sap.com/saphelp_nw04s/helpdata/en/44/9821620553053de10000000a1553f6/content.htm
    Generic Delta
    If a field exists in the extract structure of a DataSource that contains values which increase monotonously over time, you can define delta capability for this DataSource. If such a delta-relevant field exists in the extract structure, such as a timestamp, the system determines the data volume transferred in the delta method by comparing the maximum value transferred with the last load with the amount of data that has since entered the system.  Only the data that has newly arrived is transferred.
    To get the delta, generic delta management translates the update mode into a selection criterion. The selection of the request is enhanced with an interval for the delta-relevant field. The lower limit of the interval is known from the previous extraction. The upper limit is taken from the current value, such as the timestamp or the time of extraction. You can use security intervals to ensure that all data is taken into consideration in the extractions. After the data request was transferred to the extractor, and the data was extracted, the extractor then informs generic delta management that the pointer can be set to the upper limit of the previously returned interval.
    The delta for generic DataSources cannot be used with a BW system release prior to 3.0. In older SAP BW releases, the system does not replicate DataSources for master data and texts that were made delta-enabled using the delta for generic DataSources.
    Determining the Generic Delta for a DataSource
           1.      Choose Generic Delta.
           2.      In the subsequent dialog box, specify the delta-determining field and the type for this field.
           3.      Maintain the settings for the generic delta:
                                a.      Specify a security interval.
    The purpose of a security interval is to make the system take into consideration records that appear during the extraction process but which remain unextracted (since they have yet to be saved) during the next extraction.
    You have the option of adding a security interval to the upper limit/lower limit of the interval.
    A security interval should only be specified for the lower limit when the delta method results in a new status for changed records, in other words, when the status is overwritten in BW. In this case, duplicate data records that could arise in such a safety interval have no affect on BW.
                                b.      Choose the delta type for the data to be extracted.
    The delta type is used to determine how extracted data is interpreted in BW and which data targets in which it can be posted.
    With the delta type additive delta, the record to be loaded for summarizable key figures only returns the change to the key figure. The extracted data is added in BW. DataSources with this delta type can supply both ODS objects and InfoCubes with data.
    With the delta type New Status for Changed Records, every record to be loaded returns the new status for all key figures and characteristics. The values are overwritten in BW. DataSources with this delta type can write the data into ODS objects and master data tables.
           4.      Save your entries.
    Delta transfer is now possible for this DataSource.
    After generating the DataSource, you can see this from the marking for the field Delta Update on the DataSource: Customer Version screen.
    In systems from release 4.0B, you can display the current value of the delta-relevant field in the delta queue.
    Thanks and regards

  • HR Full Load error in Payroll loading from EBS

    Hi,
    We have configured HR anylatics and when runs the Full load from EBS 11.10 (Vision Demo Database) we are receiving an error which says can not allocate 128 from temp tablespace. We have found the 11i database temp has been grown hugely (more than 10GB) and hard disk space been allocated fully.
    And we found this has been occurred by payroll package (selecting from a view).
    Is there any optimization or any solution for this issue.
    Thanks.
    Nilaksha.

    and remember to fix the bug in the database first:
    UPDATE per_all_people_f
    SET effective_start_date = '07-FEB-02'
    WHERE person_id = 6272
    AND effective_start_date = '04-JAN-91'
    AND effective_end_date > '06-FEB-02';
    because you will see 2 rows if you for:
    SELECT *
    FROM per_all_people_f
    WHERE person_id = 6272
    AND effective_start_date = '04-JAN-91';
    And this isn't possible, becuase HR is date tracked and you can't start twice on 1 date!

  • Error in Transaction Data - Full Load

    Hello All,
        This is the current scenario that I am working on:
    There is a process chain which has two transaction data load (FULL LOADS) processes to the same cube.In the process monitor everything seems okay (data loads seem fine) but overall status for both loads failed due to 'Error in source system/extractor' and it says 'error in data selection'.
    Processing is set to data targets only.
    On doing a manage on the cube, I found 3 old requests that were red and NOT set to QM status red. So I set them to QM status red and Deleted them and the difference I saw was that the subsequent requests became available for Reporting.
    Now this data load which is a full load takes for ever - I dont even know why I do not see a initialize delta update option there - can Anyone tell me why I dont see that.
    And, coming to the main question, how do I get the process chain completed - will I have to repeat the data loads or what options do I have to have a succesfully running process chain or at least these 2 full loads of transaction data.
    Thank you - points will be assigned for helpful answers
    - DB
    Edited by: Darshana on Jun 6, 2008 12:01 AM
    Edited by: Darshana on Jun 6, 2008 12:05 AM

    One interesting discovery I just found in R/3, was this job log with respect to the above process chain:
    it says that the job was cancelled in R/3 because the material ledger currencies were changed.
    the process chain is for inventory management and the data load process that get cancelled are for  the job gets cancelled in the source system:
    1. Material Valuation: period ending inventories
    2. Material Valuation: prices
    The performance assistant says this but I am not sure how far can I work on the R/3 side to rectify this:
    Material ledger currencies were changed
    Diagnosis
    The currencies currently set for the material ledger and the currency types set for valuation area 6205 differ from those set at conversion of the data (production startup).
    System Response
    The system does not allow you to post transactions after changing the currency settings to ensure consistency.
    Procedure
    Replace the current settings with the those entered at production
    start-up.
    If you wish to change the currency settings, you must use programs to convert data from the old to the new currencies.
    Inform your system administrator
    Anyone knowledgable in this area please give your inputs.
    - DB

  • Load R/3 SD data in BW on delta basis in stead of full load.

    Hi,
    At present we load SD data from R/3 to BW on full load basis. But we need to change it from full load to delta load. Can anybody shed some light on the subjects.
    Thanks in advance
    Imran

    Hi,
    Welcome on board.
    There are so many threads relating to your query. Please do some search.
    There are some good blogs written by Roborto. you can find them in BLOGS.
    For sales you have application component 11,12,13 in the LBWE transaction in ECC. All the datasources are delta capable.
    Thanks
    Srikanth

  • Need to post Full Load data (55,000 records) to the target system.

    Hi All,
    We are getting the data from SAP HR system and we need to post this data to the partner system. So we configured Proxy(SAP) to File(Partner) scenario. We need to append the data of each message to the target file. Scince this is a very critical interface, we have used the dedicated queues. The scenario is working fine in D. When the interface transported to Q, they tested this interface with full load i.e with 55,000 messages.All messages are processed successfully in Integration Engine and to process in Adapter engine, it took nearly 37 hrs. We need to post all 55,000 records with in 2 hrs.
    The design of this interface is simple. We have used direct mapping and the size of each message is 1 KB. But need to append all messages to one file at the target side.We are using Advantco sFTP as receiver adapter and proxy as a sender.
    Could you please suggest a solution to process all 55,000 messages with in 2hrs.
    Thanks,
    Soumya.

    Hi Soumya,
    I understand your scenario as, HR data has be send to third party system once in a day. I guess, they are synchronizing employee (55,000) data in third party system with SAP HR data, daily.
    I would design this scenario as follows:-
    I will ask ABAPer to write a ABAP program, which run at 12:00, pickup 55,000 records from SAP HR tables and place them in one file. That file will be placed in SAP HR file system (you can see it using al11). At 12:30, PI File channel will pick up the file and transfer the file to third party target system as it is, without any transformation. File to File, pass through scenario (no ESR objects). Now, ask the target system to take the file, run their program (they should have some SQL routines). That SQL program will insert these records into target system tables.
    If 55,000 records make huge file at SAP HR system, ask ABAPer to split it into parts. PI will pick them in sequence based on file name.
    In this approach, I would ask both SAP HR (sender) and third party (target) system people to be flexible. Otherwise, I would say, it is not technically possible with current PI resources. In my opinion, PI system is middleware, not system in which huge computations can be done. If messages are coming from different systems, then collection in middleware makes sense. In your case, collecting large number of messages from single system, at high frequency is not advisable. 
    If third party target system people are not flexible, then go for File to JDBC scenario. Ask SAP HR ABAPer to split input file into more number of files (10-15, you PI system should be able to handle). At receiver JDBC, use native SQL. You need java mapping to construct, SQL statements in PI. Donu2019t convert flat file to JDBC XML structure, in your case PI cannot handle huge XML payload.
    You have to note, hardware upgrade is very difficult (you need lot of approvals depending your client process) and very costly. In my experience hardware upgrade will take 2-3 months.
    Regards,
    Raghu_Vamsee

  • CDC Error (JOURNALIZED DATA is not loading into tartet database)

    HI,
    I have enabled source database for CDC and got green mark on source database model tables.
    while inserting data into source, J$ tables updating JRN_SUBSCRIBER and other filed.
    when I run the package/ interface JOURNALIZED DATA is not loading into target database.
    i have implemented cdc for 7 source table.
    and using JKM MSSQL Simple
    and enable JOURNALIZED DATA in interface level.
    and
    source database is : MSSQL Server
    Target Database : Oralce 11g.
    please advice me.
    thanks in advance.
    Zakeer Hussain

    Zakeer look into this link , -http://odiexperts.com/?p=1096 . Hope this helps.
    also before running Can you right click on the Source Datastore and click on Journal Data and can you see the data ? and if still the data is not passing through ,in that case make temporary objects t- yes in LKM , IKM and debug and see at which step data is not flowing through and look if there is any filter or condition which is stopping it .
    Still not able to figure out please tell us which step the data is not flowing through we will try to guide you.

  • Full load from r/3 failed due to bad data and no psa option selected

    SDN Experts,
    Full load from R/3 failed loading in to ODS and in the infopackage on psa selected. how to fix it? can i rerun the infopackage again to load from R/3? May be it will fail again? is this a design defect?
    i will assign points. Thank you.
    Les

    Hi,
    There is obsolutely no prob. in re executing the package, but before that, check
    your ODS and Info Source is re active and update rules are also active if any..
    and you can check the option of target info provider selected properly along with
    the PSA option and also check whether you have any subsequent process like
    updating data from ODS to Cube and so on..
    You can also check the data availability at R/3 for the respective DS by executing
    RSA3.
    Hope this helps..
    assign points if useful..
    Cheers,
    Pattan.

  • Full load - Every month data loading without deleting previous data

    Dear SDN,
    I am loading (Full load) Data from 001.2005 to 010.2007 using 0FI_GL_1 with the selection option Fiscal Year/Period (From : 001.2005 and To : 010.2007) ...
    From next month (011.2007) onwards, when loading data everytime it deletes entire data and again loading (Since i have kept an option in Infopackage -- Datatargets Tab -- Delete entire content of datatarget)...
    My doubt is wheather without deleting the whole data, can we load from next month data onwards...so, if that is the case shall i use Fiscal year/Period range ?? (or) should not use Fiscal year/Period range...
    But the client needs data only from 001.2005 onwards not beyond that
    Please help me to resolve this issue..
    Help will be greatly appreciated with points...
    Thanks....

    Hello Venkat,
    Unfortunately you can not do that way, because in period 11.2007 (assume this is the period that you want to newly load to cube.) there might be posting in previous period. And these postings made to previous period should also reflect to cube. Only way is delete and reload by using full.
    Other option you might use a ODS before sending to cube. you might every time load full data, ODS key fields might be checked by system and neccesary inserts and updates will be done by system, so there wont be duplications. And finally only new data will be sent to cube as delta.
    Sarhan.

  • Connect to Oracle database through Excel to load data

    Dear all,
    does anybody have an idea if i can connect directly from microsoft excel to an oraclle database in order to load the data in the excel file in the database.
    Thank u

    user562674 wrote:
    Dear all,
    does anybody have an idea if i can connect directly from microsoft excel to an oraclle database in order to load the data in the excel file in the database.
    Thank uyes, EXCEL can use ODBC to connect to Oracle DB

Maybe you are looking for

  • Brain:485 Posting Period 0FISCPER3

    Hi. All, After execution of my query on the portal I am getting this error : Brain:485 Posting Period 0FISCPER3. Not sure what it means. Please help to understand this and provide me some solution. what I have done is: I just executed in RSRT also bu

  • Need help uninstalli​ng desktop and HS Mgr/Vista

    Hi, I have an HP Laptop that came with Vista, about 1 year old and updated to current.  I have Palm Desktop and Hotsync manager v6.2.2 installed and want to remove them completely.  There is no uninstall feature in the Palm folder, when I go to contr

  • Safari, App Store Issues after updating to 10.7.2

    I have been running my imac that came preinstalled with Lion I am having this wierd issue after updating to 10.7.2 recently When i run Safari, App Store, Itunes, i can see its loading.. but after 2 hours its still loading. It just seems to be stuck a

  • CSM - Is it advisable to upgrade to latest version 4.2.(4)?

    Hi Sir, I have a pair of Cisco Catalyst 6513 switches; each contains a CSM (WS-X6066-SLB-APC). Both CSMs are configured in a fault tolerant group. Both running software version 4.2.(3). Below is latest Field Notice for CSM (DDTS = CSCsd27478): http:/

  • Black areas are showing up as green

    Since I upgraded to Firefox V:8 some of the images show the black areas (background, shadows, and the like) as a bright green. If I change to the IE engine, the problem goes away. I've messed with disabling the add-ons, without any repeatable fix. So