Help in loading data from an infocube to another

all,
i have been trying to load data from a consolidation infocube of one BW system to a similar infocube in another BW system
I managed to load data by identifying the fist system as the source system
the challenge that i m facing is as follows
the first infocube has approx 70000 records; while when i uploaded the data approx only 20000 records got updated
would appreciate help on why this has happened
thanks for your time

Hi,
   Are you using Open hub interface?
   If yes, then the structure/datamodel of the two infocubes are same?
   If both the infocubes are having the same structure, then did you apply any transformation (custom)logic,selection field values in the infospoke definition ?
   If the all the above are true, then do check the the status overview of the executed infospokes grouping by logical target system, destination, infospoke and day.
   And also double click to access single request log and then drill down to record count and errors.
regards,
ravi

Similar Messages

  • How to load data from one Infocube to another request by request using DTP

    Hi All,
    I have a scenario where we are maintaining backup Infocube B for Infocube A.  User loads data from a flat file many times a day in different requests.  We need to maintain backup of the Infocube A on weekly basis.  i.e data to be refreshed with delta update from cube A to cube B.  There are some situations where user deletes some of the requests in Infocube A randomly after use.  Will this change effect in back up cube B when performed data refresh from cube A to cube B.  i.e. this functionality is similar to reconstruct in BW 3.5.  Now we are running on BI 7.0 SP 9.
    Can anyone answer this ASAP.
    Many Thanks,
    Ravi

    You cannot load request by request, " Get Data By Request " DTP loads request by request on " First In First Out " basis. You can run some Pseudo/Fake DTP's if you dont want to load data from a particular request.
    If the user deletes a request from Cube A, it wont be loaded to Cube B but if it is already loaded in Cube B and later the user deletes the request from Cube A you have to delete the request frm Cube B. Inorder to monitor request by request run DTP with   " Get Data By Request set.

  • Load data from an infocube into other infocube

    Hi guys....i try to load data from an infocube to other infocube.....i generated the infosource for export without problems i can visulize the datasource in my BW System assigned well...but how i can run the extraction from one infocube into the other?
    I view that in ODSs there's an option to run load of data into datatargets but in the infocube i can´t find this option......
    is there anything else to do?
    Regards

    The Infosurce and Datasource was linked automatically i can see this in Source System Myself but i can´t see it  in Infosources Tab and i can´t create a Infopackage....that´s my problem

  • Load data from one Application to another in BPC NW 7.5

    Hi
    I have a requirement to load data from One application to another application.
    eg. HR Plannning to CostCenter Planning.
    I am aware of a method to load data using transformation file mapping source and target dimensions and run DM package and this works fine.
    However, the challenge is if i try to automate this process then i will have to hardcode the application tech name and save as variant to use that in process chain. Now, if i hardcode the tech name of the application then it might happen that during full optimization the tech name might change and the automation load will fail.
    so how to handle this situation? any standard practices available?
    appreciate your inputs/thoughts.
    Regards.

    Hi,
    Exactly. The main problem is that there is no statement to transfer the data. Please take a look at the below link from help.sap:
    http://help.sap.com/saphelp_bpc75_nw/helpdata/en/f4/33f0bbc46c4a53860f1410f20ffb8b/content.htm
    Since the structure of both the applications is same, you dont require add_dim or rename_dim, etc. However, you definitely need the when statement.
    If you want to transfer all the records, then you can have something like
    *WHEN ACCOUNT
    *IS *
        *REC(EXPRESSION = %VALUE%)
    *ENDWHEN
    Hope this helps.

  • How load data from One Infoobject to Another Infoobject?

    Hi GuRUS
    Any body has succeded in loading data from One Infoobject to Another Infoobject?
    Pls Let me kNow
    Regards
    Durai

    Hi Durai,
    Create export datasource on the info-object and then load the data to the other info-object.
    Check this link also:
    Re: How load data from Infoobject to another Infoobject?
    Bye
    Dinesh

  • Need help in loading data from a csv file to a table in Oracle DB

    Hi,
    I am using sqlplus as a client to connect to the server.
    I am trying to load data from a csv file from a client to a table in Oracle DB server.
    I am trying to use the below command
    " LOAD DATA INFILE 'test.csv' INTO TABLE testTable FIELDS TERMINATED BY ',' (test1,test2,testc3,testc4); "
    But, I am encountered with the following error
    "SP2-0042: unknown command "load data" - rest of line ignored."
    Thanks in advance.
    SB

    Hey Frostmann,
    That was a nice post....
    I changed my mind to use an 'Insert into' statement from a shell script.
    I created a DB table named test and I tried using the below shell script to insert a row in the table.
    sqlplus test/test@test <<ENDOFSQL
    INSERT INTO test VALUES('test1',123,'test2','test3');
    exit
    ENDOFSQL
    A row is succesfully inserted into the table when I run the script manually, but it does not insert rows when a cron job is scheduled.
    Could you please help me with this?
    Thanks in advance.
    SB

  • Gurus..need help in reading data from virtual infocube

    Gurus,
    I have to read data from an virtual infocube...I am trying to use FM RSDRI_INFOPROV_READ to read data but it doesn't work..
    I am doing exactly what has been done in the demo program RSDRi_INFOPROV_READ_DEMO...
    Please help me...its really URGENT...
    Thanks
    sam

    Check out this thread....
    must be helpful
    Re: Read data from 0BWTC_C02 via ABAP

  • Quetsion about loading data from transactional infocube

    Hi,
    We have 4 transaction infocubes where the users data will be saved from their layouts and we have one more transaction infocube(A) where all these 4 transactionl data will saved to onecube say(A)
    We also load from A- to other basic cube(B).
    My questions are when
    1.I load from All these 4 trnsactional cube to A - Do I need to say transactional can be planned not loading allowed.
    2.When I load from A to b(basic cuvbe do I need to say - Planning cannot be allowed only laoding)
    Let me know
    Thanks

    If you want to load data from planning cubes to further data target using update rules or transformation, their requests need to be green (yellow request would not load further). In this case the target transactional cube should be in loading mode. If you do not want to do these things, you can look at using a copy planning function (but remember that planning function takes more time than BW loading - this difference would be really noticeable if the data volume is large).
    There are many threads on this topic - if you search, you can get more info.

  • Loading data from one table to another using cursor

    Hi,
    I have given the below command to load the data from 1 table to another using cursor.
    declare
    cursor mycursor IS
    SELECT extract_name,from_date,to_date,BETA from temp_table where EXTRACT_NAME='GIFTCARD_DETAILS';
    Begin
    for mycursor_1 IN mycursor loop
    insert into tmp_tab columns(col1,col2,col3,col5) values(mycursor_1.EXTRACT_NAME,mycursor_1.from_date,mycursor_1.to_date,mycursor_1.BETA);
    End loop;
    commit;
    end;
    It is working fine.
    But I want to hard code some of the columns ( like flags ) which are not there in 1st table and load them into 2nd table.
    In db2 we will give commands like
    varSqlStatus=`db2 "declare mycurs cursor for select extract_name,from_date,to_date,BETA,'N','Y' from temp_table"`
    varSqlStatus=`db2 "load from mycurs of cursor modified by identityignore insert into tmp_tab(col1,col2,col3,col5,col6,col7) nonrecoverable"`
    But I want it in oracle 10g, Can any one help me in this.

    Have you tried either of the two options :
    1. Modify the CURSOR itself :
    cursor mycursor IS
    SELECT extract_name,from_date,to_date,BETA,'N','Y' from temp_table where EXTRACT_NAME='GIFTCARD_DETAILS';2. Modify the INSERT statement itself :
    insert into tmp_tab columns(col1,col2,col3,col5,col6,col7) values(mycursor_1.EXTRACT_NAME,mycursor_1.from_date,mycursor_1.to_date,mycursor_1.BETA,'N','Y');

  • Data load failed while loading data from one DSO to another DSO..

    Hi,
    On SID generation data load failed while loading data  from Source DSO to Target DSO.
    Following are the error which is occuuring--
    Value "External Ref # 2421-0625511EXP  " (HEX 450078007400650072006E0061006C0020005200650066
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject 0BBP
    So, i'm  not getting  WHY in one DSO i.e Source  it got successful but in another DSO i.e. Target its got failed??
    While analyzing all i check that SIDs Generation upon Activation is ckecked in source DSO but not in Target DSO..so it is reason its got failed??
    Please explain..
    Thanks,
    Sneha

    Hi,
    I hope your data flow has been designed in such a way where the 1st DSO as a staging Device and all transformation rules and routine are maintained in between 1st to 2nd dso and sid generation upon activation maintained in 2nd DSO.  By doing so you will be getting your data 1st DSO same as your source system data since you are not doing any transformation rules and routine etc.. which helps to avoid data load failure.  
    Please analyze the following
    Have you loaded masterdata before transaction data ... if no please do it first
    go to the property of first dso and check whether there maintained sid generation up on activation (it may not be maintained I guess)
    Goto the property of 2nd Dso and check whether there maintained sid generation up on activation (It may be maintained I hope)
    this may be the reason.
    Also check whether there is any special char involvement in your transaction data (even lower case letter)
    Regards
    BVR

  • Problem in loading data from one DSO to another

    Hi All,
       I am using Multicube 0PM_MP03, this mutli cube use DSO 0PM_DS02 & 0PM_DS01, data will flow from PSA to 0PM_DS01 & then to 0PM_DS02 using a Update rule, i have triggered my infopackage & i can see data in 0PM_DS01 & i have activated the data in 0PM_DS01, now on right click on 0PM_DS01 i click on the option  Additional Functions -> Update 3.x Data to Target, now the request status is green in 0PM_DS02 but i cannot see any data in this DSO, neither in new or active, can anyone help me out in this where i am doing wrong.
    Thanks,
    Pritam.

    Hi Pritam,
    This is strange, Check the below points.
    1. If the number of records added 0, in manage of DSO2?
    2. Confirm if there is a transfprmation between ODS1 & ODS2. If there is, you need to use a DTP (not Additional Functions -> Update 3.x Data to Target).
    3. Activate the request and then check.
    4. Maintain PSA for the datasource 80PM_DS01 and when loading data from DSO1 to DSO2 choose option PSA and then into targets.
    Regards,
    Panksj.

  • Loading Data from one Cube into another Cube

    Hi Guys,
    I am trying to load data from one cube A to another cube B. Cube A has data around 200,000 records. I generate export datasource on Cube A. Replicated the datasource and created InfoSource and activated it.
    I created update rules for Cube B selecting Source as Cube A. I do have a start routine to duplicate records in Cube A. Now when I schedule load,
    It stops at Processing Datapacket and says no data. Is there something wrong with the update routine or is there any other way to load form cube to cube in a simpler way?
    Thanks in advance

    This is the start routine to duplicate records in two currencies.
    DATA: datew   TYPE /bi0/oidateto,
          datew2  TYPE rsgeneral-chavl,
          fweek   TYPE rsgeneral-chavl,
          prodhier TYPE /bi0/oiprod_hier,
          market  TYPE /bic/oima_seg,
          segment TYPE /bic/oizsegment.
    DATA: BEGIN OF S_DATA_PACK OCCURS 0.
            INCLUDE STRUCTURE /BIC/CS8ZSDREV.
    DATA: END OF S_DATA_PACK.
    S_DATA_PACK[] = DATA_PACKAGE[].
      REFRESH DATA_PACKAGE.
      LOOP AT S_DATA_PACK.
        move-corresponding s_data_pack to DATA_PACKAGE.
        if DATA_PACKAGE-loc_currcy = 'EUR'.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalgrc.
          DATA_PACKAGE-CURRENCY = 'USD'.
          APPEND DATA_PACKAGE.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalloc.
          DATA_PACKAGE-CURRENCY = 'EUR'.
          APPEND DATA_PACKAGE.
        else.
          DATA_PACKAGE-netval_inv = DATA_PACKAGE-/bic/zsdvalgrc.
          DATA_PACKAGE-CURRENCY = 'USD'.
          APPEND DATA_PACKAGE.
        endif.
      ENDLOOP.
    This is to load Quantity field
    RESULT = COMM_STRUCTURE-BILL_QTY.
    This is to load Value field
    RESULT = COMM_STRUCTURE-NETVAL_INV.
    UNIT = COMM_STRUCTURE-currency.

  • Help for loading data from flat file it is an imp for me!!

    Hi All,
    I am working on loading data. I am getting data from Informix as '|' (pipe) delimited.
    Each file consists of 20-25 columns. But I need few columns information only for this load. Thanks in advance for your reply.
    Thanks,
    RM

    go for fixed length loading. refer to some of the sample scripts
    (ref: search for *.ctl in find files )

  • Loading data from a table to another

    This question is continuation of my earlier thread
    Re: query help
    The last thread was answered by experts and was closed. But I have another problem, I'll put the whole scenario here:
    I have a table with data as
    id          status_code  emp_id
    1           a                10
    2           b                10
    3           a                20
    4           c                30
    5           k                10
    6           k                10
    7           a                10
    8           a                20
    9           z                20
    10           k                30
    11           k                30
    12           m                20
    13           k                10
    14           k                30and scripts are:
    CREATE TABLE TEST
    (id NUMBER,
    status_code VARCHAR2(5),
    emp_id NUMBER
    INSERT INTO test VALUES (1,'a',10);
    INSERT INTO test VALUES (2,'b',10);
    INSERT INTO test VALUES (3,'a',20);
    INSERT INTO test VALUES (4,'c',30);
    INSERT INTO test VALUES (5,'k',10);
    INSERT INTO test VALUES (6,'k',10);
    INSERT INTO test VALUES (7,'a',10);
    INSERT INTO test VALUES (8,'a',20);
    INSERT INTO test VALUES (9,'z',20);
    INSERT INTO test VALUES (10,'k',30);
    INSERT INTO test VALUES (11,'k',30);
    INSERT INTO test VALUES (12,'m',20);
    INSERT INTO test VALUES (13,'k',10);
    INSERT INTO test VALUES (14,'k',30);Here's the query that experts gave me in last thread, to fetch desired records:
    The requirement was to fetch records of those emp ids where a given emp_id consectively gets same status code.
    Each emp id gets a status code and at times it gets same status code as its previous status code. And I am interested in such scenarios.
    For example,
    emp_id=20 gets status_code=a in id=3 and its next status_code is also =a at id=8,
    emp_id=10 gets status_code=k in id=5 and gets same status_code again in id=6,
    emp_id=30 gets status_code=k in id=10 and gets same status code in id=11 and id=14.
    And the answer was:
    select *
    from
    select
      id,
      status_code,
      emp_id,
      lag(status_code) over (partition by emp_id order by id asc) as last_status,
      lead(status_code) over (partition by emp_id order by id asc) as next_status 
    from test
    where status_code = last_status
    or status_code = next_status
      / But now the new issue is that the records of the table TEST are to be loaded into another table TEST_REVISED . For each ID of table TEST, there will be one or more records in table TEST_REVISED.
    ALTER TABLE TEST ADD (CONSTRAINT id_pk  PRIMARY KEY(id));
    CREATE TABLE TEST_REVISED
    (TR_ID NUMBER PRIMARY KEY,
    ID     NUMBER REFERENCES TEST(ID),
    NEW_STATUS_CODE VARCHAR2(5)
    );This is what I want to see in table TEST_REVISED
    TR_ID,     ID,        NEW_STATUS_CODE
    100,        1,          'test-code'
    200,        2,'         'test-code'
    300,        3,         'test-code'
    400,        4,          'test-code'
    500,        5,          'test-code'
    600,        5,           'test-code'--notice here: EMP_ID=10 had same status consecutively in table TEST, so when that record is loaded to TEST_REVISED, ID=5 is used                instead of  id=6
    700,        7,          'test-code'
    800,        3,           'test-code'--notice here: EMP_ID=20 had same status  consecutively in table TEST, so when that record is loaded to TEST_REVISED, ID=3 is used  instead of id=8
    900,        9,           'test-code'
    1000,      10,         'test-code'
    1100,      10,           'test-code'--notice here: EMP_D=30 had same status  consecutively in table TEST, so when that record is loaded to TEST_REVISED, ID=10 is used  instead of id=11
    1200,      12,             'test-code'
    1300,      13,             'test-code'
    1400,      10,             'test-code'--notice here: EMP_ID=30 had same status  consecutively in table TEST, so when that record is loaded to TEST_REVISED, ID=10 is used  instead of id=14 or id=11Ignore the value of the column NEW_STATUS_CODE here.
    How can I load data in table TEST_REVISED from table TEST keeping in view the requirement I mentioned? Table TEST has around 1 million records.
    Once this achieved, I want to delete records from table TEST with following ids :
    ID=6,8,11,14.
    Thanks a lot!
    Edited by: user5406804 on Dec 5, 2010 5:49 PM
    Edited by: user5406804 on Dec 5, 2010 5:50 PM

    OK, i still don't see how this is all supposed to fit together.
    You haven't said it, but i'm assuming when you have a series like
    id   emp status_code
    1    30   k
    2    30   k
    3    10   a
    4    30  k
    5    30  kYou actually want the output to be
    id   emp status_code   new_id
    1    30   k                   1
    2    30   k                   1
    3    10   a                   3
    4    30  k                    4
    5    30  k                    4Which presents a bit of a problem using simply analytic functions.
    So here's the modified query that picks up just the rows you want 'changed'.
    TUBBY_TUBBZ?
    select
      id,
      emp_id,
      status_code,
      connect_by_root(id) as new_id
    from
      select
        id,
        emp_id,
        status_code,
        lag(status_code)  over (partition by emp_id, status_code order by id asc) as last_status,
        row_number()      over (partition by emp_id, status_code order by id asc) as rn
      from
        select
          id,
          emp_id,
          status_code,
          lead(status_code) over (partition by emp_id order by id asc) as next_code,
          lag(status_code) over (partition by emp_id order by id asc) as last_code,
          row_number() over (partition by emp_id order by id asc) as rn
        from test
      where status_code in (last_code, next_code)
    start with last_status is null
    connect by prior emp_id = emp_id
    and prior status_code = status_code
    and prior rn = rn - 1
         ID     EMP_ID STATU     NEW_ID
          5        10 k           5
          6        10 k           5
          3        20 a           3
          8        20 a           3
         10        30 k          10
         11        30 k          10
         14        30 k          10
    7 rows selected.
    TUBBY_TUBBZ?You say you have a lot of rows in the TEST table, and i'm hoping and praying this is a one time requirement and not an on-going process that needs to occur so your best bet (performance wise) would probably be to use the above query to create a temporary table.
    Use that temporary table to control the rest of the process (inserting into TEMP_REVISED and deleting from TEST).
    That's the best i can think of based on what i believe your requirement is ... and i honestly don't think that's going to perform overly well (the connect by functionality specifically).

  • Loading data from one table to another table

    I need to load the data (data conversion) from one table to 3 different table.
    I have to load the data from Source table to following 3 target tables.
    So Consumer table has email pk and it has relationship with CONSUMER_RCV table and CONSUMER_RCV
    table has relationship with CONSUMER_ATTR table. I am doing right now with pl/sql but I am looking
    to use MULTI INSERT or SQL*LOADER or another efficient way.
    I tried with MultiInsert but it getting failed as it's violating Unique Constraints as source table might contains
    Duplicate emails.
    Could you please show me how I can load the data efficiently?
    Also I need to load in CONSUMER_ATTR table from CONTACT table which has ATTR1, ATTR2 columns.
    so My mapping will be ...
    Contact.email, first_name, last_name will go into Consumer table
    then in CONSUMER_RCV, I generate the id column value by sequence, load email and DATE_CREATEd from contact table
    And rest of the value is hard coded
    Then in CONSUMER_ATTR table I will generate the CONSUMER_ATTR_id through sequence, for ID, will use the same
    id column value which I have generated by sequence in CONSUMER_RCV, then put the attr_type = 'ATTR1' and for
    attr_value column I will insert the value from contact table's ATTR1 value,
    Then generate another CONSUMER_ATTR_ID, will use the same id column value which I have generated by sequence
    in CONSUMER_RCV, then put the attr_type = 'ATTR2' and for
    attr_value column I will insert the value from contact table's ATTR2 value,
    Please let me know if you need further explanation.
    I am using Oracle 9i R2.
    Please consider CONTACT is Source table, CONSUMER is a Target1, CONSUMER_RCV is a Target2 and CONSUMER_ATTR is a Target3
    to simplify the problem.
    Following is the Table structure
    =========================
    CREATE TABLE CONTACT (
    ID VARCHAR2 (40) NOT NULL, -- will go into Target1
    EMAIL VARCHAR2 (100) ,      -- might be duplicate and will go into Target1 and Target2
    FIRST_NAME VARCHAR2 (100) NOT NULL, -- will go into Target1
    LAST_NAME VARCHAR2 (100) NOT NULL, -- will go into Target1
    COUNTRY VARCHAR2 (40) NOT NULL, -- will go into Target1
    PHONE VARCHAR2 (100), -- will go into Target1
    NOTIFY VARCHAR2 (3), -- will go into Target2 table's RECEIVE_EMAIL column
    CREATE_DATE DATE ,               -- will go into Target1 and target2
    ATTR1 VARCHAR2 (400),           -- will go into Target3 as attr_type= ATTR1 and will load actual value in ATTR_VALUE
    ATTR2 VARCHAR2(100),          -- will go into Target3 as attr_type= ATTR2 and will load actual value in ATTR_VALUE
    ATTR2 VARCHAR2(100),          -- will go into Target3 as attr_type= ATTR1 and will load actual value in ATTR_VALUE
    CONSTRAINT CONTACT_PK
    PRIMARY KEY ( USER_ID ) ) ;
    CREATE TABLE CONSUMER(
    EMAIL VARCHAR2 (100) NOT NULL, -- PK
    TITLE VARCHAR2 (40),
    FIRST_NAME VARCHAR2 (40),
    LAST_NAME VARCHAR2 (40),
    ADDRESS1 VARCHAR2 (40),
    ADDRESS2 VARCHAR2 (40),
    CITY VARCHAR2 (30),
    STATE          VARCHAR2 (30),
    ZIP          VARCHAR2 (10),
    COUNTRY VARCHAR2 (40),
    PHONE VARCHAR2 (15),
    DATE_CREATED DATE NOT NULL,
    CONSTRAINT CONSUMER_PK
    PRIMARY KEY ( EMAIL ) ) ;
    CREATE TABLE CONSUMER_RCV (
    ID                VARCHAR2 (40) NOT NULL,-- PK
    EMAIL VARCHAR2 (100) NOT NULL,-- FK reference to Consumer
    SITE VARCHAR2 (100) NOT NULL, -- default website
    RECEIVE_EMAIL VARCHAR2 (1),
    CREATE_DATE DATE NOT NULL,
    CONSTRAINT CONSUMER_RCV_PK
    PRIMARY KEY (ID) ) ;
    ALTER TABLE CONSUMER_RCV ADD CONSTRAINT CONSUMER_RCV_FK
    FOREIGN KEY (EMAIL)
    REFERENCES CONSUMER (EMAIL) ;
    CREATE TABLE CONSUMER_ATTR (
    CONSUMER_ATTR_ID           VARCHAR2 (40) NOT NULL, -- PK
    ID                     VARCHAR2 (40) NOT NULL, -- FK reference to COnsumer_RCV
    ATTR_TYPE           VARCHAR2 (100) NOT NULL,
    ATTR_VALUE           VARCHAR2 (4000) NOT NULL,
    CONSTRAINT CONSUMER_ATTR_PK
    PRIMARY KEY ( CONSUMER_ATTR_ID ) ) ;
    ALTER TABLE CONSUMER_ATTR ADD CONSTRAINT CONSUMER_ATTR_FK
    FOREIGN KEY (CONSUMER_ATTR_ID)
    REFERENCES CONSUMER_RCV (ID) ;

    HI Hema,
      How are the entries related.. is it like for one entry in y table there are more than one entry in x table then you have to use loop with in loop , here are the both the conditions.
    1) For each entry in Y there are more than one entries in X
      sort y by quota trpid.
      sort x by quota trpid.
      loop at y.
      loop at x where quota eq y-quota
                            and trpid eq y-trpid.
      z-value = x-value.
      append z.
      endloop.
      endloop.
    2) For each y there is one entry in x table.
    sort y by quota trpid.
      sort x by quota trpid.
      loop at y.
      read table x with key quota = y-quota
                                trpid = y-trpid binary search.
    if sy-subrc eq 0.
      z-value = x-value.
      append z.
    Endif.
      endloop.
    Mahesh

Maybe you are looking for

  • Schedule line confirmed but unable to create delivery

    hi, schedule lines are confirmed but still not able to create delivery, stock is available in the storage location, no reservations are against the storage location, when i try to create a delivery, delivery quantity field is empty then when i try to

  • Oracle db 11g r2 Configuration of processes and session

    I recently installed Oracle DB 11g R2 and 50,000 users connect to this database suggest me what would be basic configurations . Processes and sessions parameters for database are 5000 and 5505 but it is little lagging now so if its possible i want to

  • OLAP enabling in the DB

    I'm not sure if OLAP was installed in my DB when i do execute SQL> select parameter, value from v$option where parameter='OLAP'; PARAMETER VALUE OLAP FALSE how do i make this true and enable the OLAP option. thanks seshan

  • Flushing the DNS cache

    I'm having trouble with a Web site when I access it on my home computer, yet this same site looks fine on my Mini at work. One section of the index page generates a "can't find server" error, and the site's own logo won't display properly. It is a fr

  • Application Server Error (dependencies)...

    Hi everyone, we have a problem everytime when we restart the Oracle Application Server (OAS). It looks like the OAS components didn't start and we get the next error: Check the exception thrown by the Repository API for resolution information. Some c