New table contains data after Successful activation

HI All,
one DSO activation got failed due to red request which was present in the target. We repeated the DSO Activation step once we deleted the bad request from DSO and it got completed successfully.
Generally, as part of ODS Activation, the data will move to ACTIVE table and CHANGELOG table and after that the same data will get delete from the ACTIVATION(NEW) queue (u201CU tableu201D).
After that DSO Activation termination, U table data is not getting deleted for that particular table.
As per the SAP note u201C680480u201D, if any termination happens while activating DSO then there might be a chance of just activating the request after that it wonu2019t delete the activation queue. And also, from that point of time onwards the new table continue to grow.
The above mentioned sap note contains solution upto the version BW 3.5. As my system is BI 7.0, we cant implement any patches as mentioned in the note.
Can any one please tell me in which table and all do i need to delete the entry of that particular request from the table apart from RSODSACTREQ.
Reards,
Sridevi.

Hi Sridevi,
Actually it's not advisable to delete the entries from  the DB tables. But at times, we are forced to do that to avoid inconsistencies in the system.
If there are no pending requests in DSO for activation and if you are able to upload further to downsteam data targets like Cubes, you need not to worry about Activation Queue.
For the time being, since you have already deleted bad entries from RSODSACTREQ and RSREQICODS tables, I feel there will not be any inconsistencies in the system.. So do not delete from any more tables.
In the next complete data load, if any inconsistency found, then you delete from the other tables.
In order to avoid this in future, make the status to red in request monitor before deleting from  the DSO Manage.
Regards,
Suman

Similar Messages

  • How-To "Refresh a table of data after inserting or deleting"

    I'd like to say a word on the how-to article "How to refresh a table of data after inserting or deleting a row using ADF".
    (http://www.oracle.com/technology/products/jdev/howtos/1013/updtable/index.html?_template=/ocom/technology/content/print)
    I spent a lot of time on it because I needed help in implementing simple CRUD functionality on a table, using JSF-ADF-TopLink technologies.
    While the the article does provide correct steps, it is in one important place not specific enough, so the reader may easily get stuck. In section "Refresh the data table", point 1: when you double click on the removeEntity() button, in Structure window, then you do not get the required dialog. You get CommandButton Properties dialog.
    You must click on the removeEntity() button in Editor's Design view. But even there you may get the CommandButton Properties dialog, not managed beans dialog.
    You may resolve that by going to JSF configuration file, faces-config.xml, and switch to Overview view. This will show you the managed beans that you have.
    Then, you may already have a backing bean for the page. You can use that and avoid creating a new managed bean.
    I could understand what the operations mean only after very careful reading of "Creating More Complex Pages", section "Overriding Declarative Methods" in JDeveloper Help (or in ADF Developer's Guide PDF document).
    In general: I believe that "ADF bindings" need more conceptual explanation, maybe in form of an article. Grammatical form "bindings" may create a false understanding that "bindings" are just references. But they are not -- ADF bindings are active objects that handle traffic between UI components and Data Controls. It seems that "bindings" even communicate among themselves. Maybe it would be more understandable to differentiate strictly between "binding objects" (or "binders"?), binding object definitions and binding object references.
    It would be very helpful to have a diagram showing grahically what specific binder objects are created in a small apllication (2-3 pages using 1-2 tables), with whom they communicate and what type of data is passed on.
    Priit

    Hi
    Thanks for your infos.
    Yes exactly I use almost the same code you have post here.
    Could You answer to my next questions?
    First - >what do you mean by saying that "it's not good idea using refreshing in IE?" Of course I use refreshing in backing_bean for my button "remove" that removes row, commit changes to database and refresh table, almost the same as You said in your post:
    Code in backing_bean is and comments on difference to Your code is below:
    public commandButton2_action1(){
    BindingContainer bindings = getBindings();
    OperationBinding operationBinding =
    bindings.getOperationBinding("removeEntity");
    Object result = operationBinding.execute();
    if (!operationBinding.getErrors().isEmpty()) {
    return null;
    //above remove entity, but I dont now if it do commit to database? So i do it below
    OperationBinding commit1 = bindings.getOperationBinding("Commit");
    commit1.execute();
    //and at the end I refresh my table, "findAllRezerwacja1 - it is an id of the methodAction, not the iterator -> is it ok? or should I change to Iterator id?
    OperationBinding requery = bindings.getOperationBinding("findAllRezerwacja1");
    requery.execute();
    return null;
    Page Definition code for this:
    <methodAction id="findAllRezerwacja1"
    InstanceName="SessionEJBLocal.dataProvider"
    DataControl="SessionEJBLocal" MethodName="findAllRezerwacja"
    RequiresUpdateModel="true" Action="999"
    ReturnName="SessionEJBLocal.methodResults.SessionEJBLocal_dataProvider_findAllRezerwacja_result"/>
    <table id="findAllRezerwacja2" IterBinding="findAllRezerwacja1Iter">
    <AttrNames>
    <Item Value="dataDo"/>
    <Item Value="dataOd"/>
    <Item Value="idRezerwacja"/>
    <Item Value="liczbaUczestnikow"/>
    <Item Value="prowadzacy"/>
    <Item Value="uwagi"/>
    </AttrNames>
    </table>
    <methodAction id="removeEntity" InstanceName="SessionEJBLocal.dataProvider"
    DataControl="SessionEJBLocal" MethodName="removeEntity"
    RequiresUpdateModel="true" Action="999">
    <NamedData NDName="entity"
    NDValue="${bindings.findAllRezerwacja2.currentRow.dataProvider}"
    NDType="java.lang.Object"/>
    </methodAction>
    <action id="Commit" IterBinding="findAllRezerwacja1Iter"
    InstanceName="SessionEJBLocal.dataProvider"
    DataControl="SessionEJBLocal" RequiresUpdateModel="true"
    Action="2"/>
    </bindings>
    //and rest of code for Iterator etc
    My second question is, why when you use refresh button in IE (I know is not recommended as You said, but sometimes user do it, so I want prevent situations that I will describe here) so when I press refresh button in IE exactly after removing one row by clicking my button, refreshing by pressing IE button is doing the same --> is deleting next row. How to stop deleting row, when for example user would press IE refresh button after pressing remove button for table. If I change selection in table after deleting row, and press refresh button in IE, instead of deleting row, I got error message: JBO-29000: JBO-35007: and
    JBO-35007. So where Im doing wrong. Maybe I should do sth with postback ? Could You help me? Thanks in advance
    Last one question: what is the difference between using delete and removeEntity from operations node? Im now reading carefully ADF Dev Guide, so I hope I can find infos there? But if You know, please answer to this question.
    Thanks

  • [Solved] UEFI. System hangs on boot after successfully activating swap

    Alright.
    Unnecessary Background:
    I've been using Arch for years. I just bought a stupidly good computer with a bunch of specs that I don't think are relevant, except for the fact that I am using an X99-pro. It has RAID support, but I am not using that currently. I bought all the parts and finished constructing it this morning (I live in Japan, so that was a while ago). Normally, I have no problems with installation; however, I wanted to do this one UEFI mode. After booting into the live disk, I checked to make sure everything was working as it should. Arch recognized my graphics card, most of my ram (one of the cards is a bit finicky for some reason), my SSD and HDD, along with the built-in WiFi (although it is an annoying broadcom one, so I need to install the driver separately later). Now, there are plenty of sources of error in here:
    Problem:
    I got through the arch installation and tried gummiboot. It would hang every time after successfully activating swap. I tried grub, too, but had the same problem. Because gummiboot is much less cumbersome, I switched back to that. Interestingly, systemd will sometimes recognize my drive twice before activating swap, and if I edit my fstab to ignore the swap completely, it hangs on the next step with is mounting the boot partition. I would like to point out that because I am not able to successfully boot my system, I cannot provide the exact messages for you guys, which may pose a problem.
    Solutions I have tried:
    1. Chrooting / mkinitcpio -p linux
    2. modifying fstab / removing swap from it completely
    3. fresh re-install
    Sources of error:
    1. Construction. Everything seems appropriate on the live usb, but maybe there is a problem in which ports I used for the SSD and HDD
    2. UEFI. It could be that UEFI just doesn't like mounting multiple things and freezes up when put on the spot.
    3. Legacy. It could be that this is not a problem with UEFI or my construction, but something I missed in the installation.
    Additional info:
    Partitioning scheme:
    sda1: ~500MiB boot / EFI partition
    sda2: ~4G swap
    sda3: ~110G /
    sdb:   ~2T, untouched (for storage later)
    I apologize for any typos and hope it was readable. I am super exhausted and going to take a nap. I'll check back in on this in the morning.
    Last edited by Leios (2015-02-08 00:16:27)

    Please post the output of:
    # gdisk -l /dev/sda
    lsblk -f
    And the contents of your gummiboot loader.conf and arch.conf (or whatever you've called it).
    Post the output of the commands after you have mounted all your partitions and (arch-)chroot'd into your system.
    You can use wgetpaste to post the outputs, as described by @ewaller here:
    https://bbs.archlinux.org/viewtopic.php … 8#p1500528
    Alternatively, pipe the command output to a text file and export it (using a live distribution or a USB stick).
    You should remove swap from your fstab -- systemd mounts it automatically anyway.

  • Can I access WorkFlow Container Data After Archiving?

    Hello .....
    I am tasked to Archive thousands of old WorkFlows that are cluttering up our system and are beginning to cause performance issues.
    I have managed to Archive some WorkFlows in our DEV system with Tcode SARA.
    My question is .... Can I still access container data for Archived WF's?  I can see logs that indicate successful Archive but our Managers still want to be able to see the container data.  is it possible or is it gone?  tahnk you in advance.
    Dan Avis

    Hello,
    Your archiving isn't successful until you've tested it out - and answered your own question.
    But as far as I can recall (I'll check) - yes, the container values are present. So far I haven't been able to find anything that was missing.
    regards
    Rick Bakker
    hanabi technology

  • Join multiple tables to create one new table with data from all the other

    CREATE TABLE NAME2012 (STOCKNAME2012 VARCHAR(255), VAL2012 FLOAT)
    INSERT INTO NAME2012 VALUES (‘Stock3’, 50)
    INSERT INTO NAME2012 VALUES (‘Stock2’, 55)
    INSERT INTO NAME2012 VALUES (‘Stock4’, 45)
    CREATE TABLE NAME2011 (STOCKNAME2011 VARCHAR(255), VAL2011 FLOAT)
    INSERT INTO NAME2011 VALUES (‘Stock1’, 56)
    INSERT INTO NAME2011 VALUES (‘Stock2’, 48)
    INSERT INTO NAME2010 VALUES (‘Stock3’, 59)
    CREATE TABLE NAME2010 (STOCKNAME2010 VARCHAR(255), VAL2010 FLOAT )
    INSERT INTO NAME2010 VALUES (‘Stock1’, 47)
    INSERT INTO NAME2010 VALUES (‘Stock2’, 46)
    I would like a new Table STOCKS populated with names of stocks from the other tables as well as values. So I get a new table STOCKS, with data as below.
    CREATE TABLE STOCKS (NAME VARCHAR(255), VALUE2012 FLOAT , VALUE2011 FLOAT, VALUE2010 FLOAT)
    End result in table STOCKS should be like:
    NAME               VALUE2010              VALUE2011             
    VALUE2012
    Stock1                 47                          
    56                            NULL
    Stock2                 46                          
    48                             50
    Stock3                NULL                         59                          
     55
    Stock4                NULL                       NULL                         
    45
    How can I write a T-SQL expression that could do this?

    Hi
    There must be StockMaster
    Try this
    CREATE TABLE NAME2012 (
    STOCKNAME2012 VARCHAR(255)
    ,VAL2012 FLOAT
    INSERT INTO NAME2012
    VALUES (
    'Stock3'
    ,50
    INSERT INTO NAME2012
    VALUES (
    'Stock2'
    ,55
    INSERT INTO NAME2012
    VALUES (
    'Stock4'
    ,45
    CREATE TABLE NAME2011 (
    STOCKNAME2011 VARCHAR(255)
    ,VAL2011 FLOAT
    INSERT INTO NAME2011
    VALUES (
    'Stock1'
    ,56
    INSERT INTO NAME2011
    VALUES (
    'Stock2'
    ,48
    CREATE TABLE NAME2010 (
    STOCKNAME2010 VARCHAR(255)
    ,VAL2010 FLOAT
    INSERT INTO NAME2010
    VALUES (
    'Stock1'
    ,47
    INSERT INTO NAME2010
    VALUES (
    'Stock2'
    ,46
    INSERT INTO NAME2010
    VALUES (
    'Stock3'
    ,59
    Create table stockMaster(STOCKNAME VARCHAR(255))
    go
    Insert into stockMaster values('Stock1')
    Insert into stockMaster values('Stock2')
    Insert into stockMaster values('Stock3')
    Insert into stockMaster values('Stock4')
    Select Distinct S.STOCKNAME , A.VAL2010 ,B.VAL2011,C.VAL2012 from  stockMaster S
    left join NAME2012  C on S.STOCKNAME = C.STOCKNAME2012
    left join NAME2011 B on s.STOCKNAME = B.STOCKNAME2011
    left join NAME2010 A on s.STOCKNAME = A.STOCKNAME2010
    go
    Drop table NAME2010
    Drop table NAME2011
    Drop table NAME2012
    Drop table stockMaster 
    Mark as Ans if you find it useful
    Shridhar J Joshi Thanks a lot

  • AR - New estimated payment date after due date

    Dear friends,
    In the event the Client is not paying within the expected due date that was calculated as "Net Due Date", we normally give the Client a grace period, and the need is to record this new estimated payment date in a date field, that the SAP user will manually introduce. The Net Due Date must not be affected as in case of having debt interests, they will be calculated on that basis. But I can not find a suitable way to implement it.
    Regards,
    Érika Iranzo
    (PS: Extra information: We calculate the "Net Due Date" (FAEDT) of the our clients and vendor's invoices from the "Baseline Date for payment" (ZFBDT) plus the highest number of days within the terms of payment "Cash discount days 1" (ZBD1T). So that the due date is calculated as FAEDT = ZFBDT + ZBD1T. Additionally we also consider a region-specific calendar)
    Edited by: EIranzo on May 14, 2009 10:14 AM

    I found the field Grace Days RFOPS-TOLER that seems to be useful for my purpose, does any one know how can I make it visible when editing a document?
    Thanks,
    EIranzo

  • Why do dimension tables contain data?

    Hi there,
    as far as I understodd the BW datamodel, the dimensions are to summon characteristics that do belong to the specific view I want to access via my InfoCube. So, the dimensional tables should contain just the mapping between DimIDs and SID, shouldn't they? What else do these dimensional tables contain??
    Thanks,
    Pascal

    Hi,
    I think the concept of using dimension table as the intermediate to SID And Facttable is to develope <i>extended star schema</i>. So that it allowing more than 13 characterstics in infocube.
    With rgds,
    Anil Kumar Sharma .P

  • What tables contain data for the CUP requests in 5.3?

    What are all of the tables that contain the data that appears in the CUP requests?  We are on version 5.3 SP13.
    We are selling off one of the divisions of our company and one of the terms is that we have to provide all relevant data, including CUP requests.  Since there are several hundreds of requests for this division (last count was over 600), it is not practical to just download the individual requests out of CUP.  So plan B is to just give them the data.
    I know there are several tables that contain this data, and I know some of them (such as VIRSA_AE_REQD_HDR and VIRSA_AE_RQD_WPHS), but I don't know all of them.  (and I would rather not have go thru and to check every table)
    Thanks.

    Hi Bob,
    I have never looked at the VT_AE tables to extract any information. Since your requirement is unique, I have a positive note that SAP would help you with it. Giving a try would worth it sometimes
    Else, paste the list of tables here, so that some one can help you.
    Have a great weekend!!
    Cheers,
    Raghu

  • System table contains data types

    Hi all,
    I am migrating queries from SqlServer 7.0 to Oracle. I found a system tabel Systypes in SqlServer 7.0 which contains all the avaliable datatypes. Is there any similar table in Oracle from which we can get all data types. Any reply will be appreciated.

    Not only at least in 10g R2:
    SQL> select type_name from dba_types where predefined = 'YES';
    TYPE_NAME
    KOKED
    KOKED1
    KOTAD
    KOTADX
    KOTMD
    KOTMI
    KOTTB
    KOTTBX
    KOTTD
    BFILE
    BINARY ROWID
    TYPE_NAME
    BINARY_DOUBLE
    BINARY_FLOAT
    BLOB
    CANONICAL
    CFILE
    CHAR
    CLOB
    CONTIGUOUS ARRAY
    DATE
    DECIMAL
    DOUBLE PRECISION
    TYPE_NAME
    FLOAT
    INTEGER
    INTERVAL DAY TO SECOND
    INTERVAL YEAR TO MONTH
    LOB POINTER
    NAMED COLLECTION
    NAMED OBJECT
    NUMBER
    OCTET
    OID
    PL/SQL BINARY INTEGER
    TYPE_NAME
    PL/SQL BOOLEAN
    PL/SQL COLLECTION
    PL/SQL LONG
    PL/SQL LONG RAW
    PL/SQL NATURAL
    PL/SQL NATURALN
    PL/SQL PLS INTEGER
    PL/SQL POSITIVE
    PL/SQL POSITIVEN
    PL/SQL RECORD
    PL/SQL REF CURSOR
    TYPE_NAME
    PL/SQL ROWID
    PL/SQL STRING
    POINTER
    RAW
    REAL
    REF
    SIGNED BINARY INTEGER(16)
    SIGNED BINARY INTEGER(32)
    SIGNED BINARY INTEGER(8)
    SMALLINT
    TABLE
    TYPE_NAME
    TIME
    TIME WITH TZ
    TIMESTAMP
    TIMESTAMP WITH LOCAL TZ
    TIMESTAMP WITH TZ
    UNSIGNED BINARY INTEGER(16)
    UNSIGNED BINARY INTEGER(32)
    UNSIGNED BINARY INTEGER(8)
    UROWID
    VARCHAR
    VARCHAR2
    TYPE_NAME
    VARYING ARRAY
    67 rows selected.

  • How to refresh a table of data after inserting a row using ADF BC

    Hi,
    i am using ADF BC, JSF to construct a page to browse and another to create/update record. But after creating or updating the record it is not reflected when i come back to browse page.
    I tried to insert invokeaction in page definition page but i can not find the iterator bindings for that table in Binds option.
    can any one provide any example how to do it.
    Please help me.

    On the page def for your submit page, add inside the executables (on the structure pane) your iterator you want refreshed (if it isn't already there).
    Then double click on the submit button and bind it to a backing bean. My method in the backing bean looks like:
        if ( !operationBinding.getErrors().isEmpty() )
            return null;
        DCBindingContainer dcbindings = (DCBindingContainer)bindings;
        DCIteratorBinding appsIter = dcbindings.findIteratorBinding("ApplicationsIterator");
        appsIter.executeQuery();
        return null;
    }The executeQuery method will refresh that iterator for your return page. The "ApplicationsIterator" is what the structure pane calls your iterator. Works for me.
    Brian

  • Date format accepted in new table but not in active table for DSO, BI 7.0

    Hi All,
    Need help from the experts, i was trying to load Flate file data in CSV(Coma separated) and cell type general for dates(yyyymmdd). everything was fine i have a routine for new fields in DSO for Fiscal week and Fiscal Quater which is calculating based on my Flate file date i.e delivery date(yyyymmdd).
    The New table of DSO means before activation everything was fine and all the logics and mappings working for the Fiscal week and Fiscal Quater and that can be visible, But after activation the same only the Fiscal week and Fiscal Quater is not showing data, its comming as blank and '000000' for all records.
    As the Fiscal week and Fiscal Quater is maintained as custmised InfoObject with data type NUM.
    Can you tell what i am missing, because i tried with different cell type also in CSV File.
    Thanks and Regards,
    Taps
    Edited by: Taps on Oct 1, 2010 1:21 PM

    Resolved...!!!
    Follow the link ...
    /people/maheshsingh.mony5/blog/2010/09/24/update-behavior-of-end-routine-in-transformations
    Regards,
    Taps
    Edited by: Taps on Oct 1, 2010 2:02 PM

  • Incorrect data after activating the request through Process chain.

    Dear SDN chaps.
    Today morning. I encountered a strange issue in DSO..
    I have DSO which is updating from the AL11(application server) flat file.
    While i am loading it to PSA there were no issues and after loading it to the DSO there is no issue and its passing through the routine and the data is populating properly in NEW data Table .But after successful activation of  the request through process i am getting the wrong records in active data table.
    Then i deleted the request and reran it manually i mean triggered the DTP and ran the manual activation surprisingly accurate records are coming through manual process..
    I am just wondering why it is not working through process chain and why it is showing incorrect records through process chain execution and how it is showing accurate records through manual uploading process..'
    Could some one please help to come out from this..By the way mine is SAP BI 7 SP20 &SP05 for BW 7.01
    Thanks
    K M R
      

    Hi Pra
    Thanks for your response..
    We are doing PSA deletion and then we are uploading the data to PSA as well as DSO.
    Now the issue is not in the part of loading we are facing the issue in Actiation of request if i am executing the activation through process chain it is sucess but the values are incorrect. If i am doing the manual activation it sucess with correct data.
    Even i tried with a new chain but still i am facing the issue.
    Surprise thing is in new data table the data is perfect in both the ways like manual upate and Process chain update only during activation i am getting incorrect record in the active data table..
    Appreciate your help on this....
    Thanks
    K M R
    Edited by: K M R on Jul 9, 2010 11:09 AM

  • Creat new table from multiple tables in same DB?

    Trying to create a new table bringing in all the records from
    3 old tables. All three old tables have common foreign key fields.
    I'm not sure how the syntax would look like, but something
    like:
    <cfquery name="qCreateTable" datasource="somedb">
    CREATE TABLE newtable
    AS
    SELECT old_ID as new_ID,
    FROM oldtable1, oldtable2, oldtable3 etc.</cfquery
    Is this possible in one step, or would I need to create the
    new table with data from oldtable1, then update with data from 2, 3
    etc.? BTW I already tried that with no success (syntax errors).
    For now, this is in Access.
    Thanks for your help.
    rinorman

    Phil:
    <cfquery name="qCreateTable" datasource="somedb">
    CREATE TABLE vendor_new2
    AS
    SELECT
    vendor_ID as old_vendor_ID,
    company_name as vendor_company,
    DBA_name as vendor_dba,
    web_URL as vendor_url,
    AdminID as old_admin_ID,
    Comment as old_comment,
    BillingLocationID as old_billing_ID
    FROM vendor_old
    UNION
    SELECT
    UserID as old_user_ID,
    EmailAddress as vendor_email,
    Password as vendor_pw,
    LastName as contact_lname,
    FirstName as contact_fname,
    PhoneNum as contact_tele,
    PhoneExt as vendor_phonex,
    FaxNum as contact_fax
    FROM main_contact
    UNION
    SELECT
    PhysicalLocationID as old_location_ID,
    Street1 as vendor_address1,
    Street2 as vendor_address2,
    City as vendor_city,
    StateID as vendor_state,
    Zip as vendor_zip,
    Zip4 as vendor_zip4,
    Phone as vendor_tele,
    Fax as vendor_fax
    FROM PhysicalLocation
    </cfquery>
    Result: Syntax error in CREATE TABLE statement.

  • Reg: Fact table and Dimension table in Data Warehousing -

    Hi Experts,
    I'm not exactly getting the difference between the criteria which decide how to create a Fact table and Dimension table.
    This link http://stackoverflow.com/questions/9362854/database-fact-table-and-dimension-table states :
    Fact table contains data that can be aggregate.
    Measures are aggregated data expressions (e. Sum of costs, Count of calls, ...)
    Dimension contains data that is use to generate groups and filters.
    This's fine but how does one decide which columns to consider for Fact table and which columns for Dimension table?
    Any help is much appreciated.
    Pardon me if this's not the correct place for this question. My first question in the new forum.
    Thanks and Regards,
    Ranit Biswas

    ranitB wrote:
    But my main doubt was - what is the criteria to differentiate between columns for Fact tables and Dimension tables? How can one decide upon the design?
    Columns of a fact table will often be 'scalar' attributes of the 'fact' data item. A dimension table will often be 'compound' attributes of a 'fact'.
    Consider employee information. The EMPLOYEE table can be a fact table. It might have scalar attribute columns such as: DATE_HIRED, STATUS, EMPLOYEE_ID, and so on.
    Other related information that can't be specified as a single attribute value would often be stored in a 'dimension' table: ADDRESS, PHONE_NUMBER.
    Each address requires several columns to define it: ADDRESS1, ADDRESS2, CITY, STATE, ZIP, COUNTRY. And an employee might have several addresses: WORK_ADDRESS, HOME_ADDRESS. That address info would be stored in a 'dimension' table and only the primary key value of the address record would be stored in the EMPLOYEE 'fact' table.
    Same with PHONE_NUMBER. Several columns are required to define a phone number and each employee might have several of them. The dimension tables are used to help 'normalize' the data in the employee 'fact' table.
    And that EMPLOYEE table might also be a DIMENSION table for other FACT tables. A DEVELOPER table might have an EMPLOYEE_ID column with a value that points to a 'dimension' row in the EMPLOYEE dimension table.

  • Data vanishing in new fields after request activation in ODS

    Hi,
    I have added some new fields in ODS and updated it by end routine .Now when i run the dtp and check the data by pressing NEW DATA tab then i am able to see data updated in the new fields.
                       When i tried to activate that request in ODS ,the values that i populated in new fields got vansihed .I am not able to find wats d issue after activation even though before activating request in ODS all data is perfectly updated in new data table.
    Regards,
    Montz

    I have faced the same issue.Data was coming till new table,but after activation fields were getting
    blank.
    There is no necessity for doing a dummy mapping. In the transformation screen you will be having an ikon near to the delete
    routine symbol with an arrow symbol
    similar to select all and unselect all, just clikc on it and then mention the setting as Update all the target fields need to be populated.
    This will solve the issue, if it is not solved, then you can make a direct mapping to any of the field in the source, it need not be the
    exact field in the target.
    Thanks,
    Saveen Kumar

Maybe you are looking for