Tables in layman laung.

Hi all,
can we revise the  config tables like VKOAand OBYC from FI and other tables from other modules to be recived in laymen lang. to provide the same to client for his better understanding.
vinila

So what do you mean by revise?
And what you want to provide to the client?
Be clear and specific with your question,
Regards
Karthik D

Similar Messages

  • Qualified Tables Working

    Hi All,
    How to Map the qualified fields in the qualified tables if the qualifier for that field is set to YES, because such fields are not reflected in the<b> import manager </b> for mapping only the fields whose qualifier is set to NO are visible.Thee problem is wen the qualifier for a field is set to YES you can feed the data only in the DATA MANAGER as these fields are only visible therewhich  become so much of a lengthy process if data is large.
    Can any one please explain the importance of the qualifier in the qualified table in layman terms, how to make the best use of it and please if possible give a short and sweet example for it.
    HELP NEEDED URGENTLY
    Thanxs in Advance.

    Hi Prasad,
    <b>Where do we use the Qualified tables:</b>
    At times data is stored in such a way that duplication is unavoidable due to the storing mechanism and other factors. It may also happen that the data is sparse. The efficient way of storing data in such scenarios is the use of Qualified tables as it reduces the size of the main table and removes the unnecessarily created duplicates.
    <b>Difference between Qualified lookup table and Flat lookup table:</b>Flat lookup tables normally work on only a single field. It contains the set of legal values to which corresponding lookup field in the main table is assigned.
    E.g. suppose we have a flat lookup table for Material Group for an automobile manufacturing company. Here we have a fixed set of legal values that can be looked up into the main table (for example Cars, Bikes, SUVs, etc.)
    Qualified tables work on the combination of two types of fields where value of one or more fields (qualifiers) changes depending on the values of one or more other fields (non-qualifiers).
    E.g. suppose we have a field “Price” whose values varies with the region for the same product. Here we have a fixed set of related legal values (combination of fields) that can be looked up into the main table. (For example Product A has a price of $30 in Southern region whereas $40 in Central region, then we get a set like Product A | Southern | $30, Product A | Central | $40 in the Lookup Qualified table)
    <b>Watch out for the definition of Qualifiers and non-Qualifiers:</b>
    This was something which took a long time for me to hunt down.
    Qualifiers – are those fields whose values change based on the value of some other field(s) and whose value is different for each main table record
    Non Qualifiers – These fields are only the part of qualified lookup table, but they apply not only to the qualified table but also to each association of the qualified table record to the main table record. Simply speaking, they are the fields that will decide the values in the Qualifier (main table)fields.
    <b>Our Scenario:</b>
    Suppose we have been provided with the following source data –
    Manufacturer Name of Drug Region Price Storage Temp
    A Crocin North 280 27 Celsius
    A Crocin South 300 14 Celsius
    A Crocin Central 260 30 Celsius
    A Disprin North 380 30 Celsius
    A Disprin South 360 27 Celsius
    A Disprin Central 320 14 Celsius
    <b>Observation:</b>
    Values of ‘Price’ and the ‘Storage Temp’ field change based upon the ‘Name of Drug’ and ‘Region’ field values and also their values are different for each main table record; hence these should be madequalifiers in the qualified lookup table
    Since ‘Name of Drug’ and ‘Region’ fields are deciding the qualifier values hence these should be kept as non qualifiers in the qualified lookup table. ‘Manufacturer’ and ‘Price’ will be part of the main table.
    <b>Design of repository:</b>
    We will discuss only the design of the main table ‘Products’ and the Qualified lookup table ‘Prices’ as they are of our immediate concern.
    Products
    Field Field Detail
    Manufacturer Name- Manufacturer
    Type - Text
    Required – Yes
    Display – Yes
    Lookup[Price] Name- Lookup[Price]
    Type - Lookup[Qualified Flat] (multi-valued)
    Required – None
    Display – No
    Multi-Valued - Yes
    Lookup Table – Prices
    Prices
    Field Field Detail
    Name of Drug Name- Name of Drug
    Type - Text
    Required – No
    Display – Yes
    Qualifier - No
    Region Name- Region
    Type - Text
    Required – No
    Display – Yes
    Qualifier - No
    Price Name- Price
    Type - Currency
    Required – No
    Display – Yes
    Symbol - $
    Decimal Place - 0
    Qualifier - Yes
    Cache - Yes
    Storage Temp Name- Storage Temp
    Type - Measurement
    Required – No
    Display – Yes
    Dimension – Temperature
    Default Unit – Celsius
    Qualifier - Yes
    Cache - Yes
    Note: Caching of qualifiers is strongly recommended as it dramatically improves search performance.
    Once the design of the repository is through, we now move on to the trickiest area – the one of the Import Manager
    Importing records:
    As per the standard practice, we load the lookup table first, here ‘Prices’ and then the main table ‘Products’.
    <b>Step1:</b>
    Loading the Qualified lookup table data:
    Select the corresponding tables in the source and destination hierarchy in the Import manager.
    Go to Map Fields/Values tab and Map the corresponding fields in the field mapping grid.
    Note: Here only two fields (non qualifiers)can b seen in the destination field grid, however in the repository design we have created four fields.
    The qualifiers (fields) never appear in the destination field of the qualified lookup table in the Import Manager.
    Go to Match Records tab and select the non-qualifiers as matching field and Add it as Combination.
    Select Import Action as Create as it is an Initial load of data.
    <b>
    Step2:</b>
    Loading the Main table data:
    Map all the source fields to corresponding fields in the repository.
    Notes: Here many fields in destination (including qualifiers (Q) which are defined in the qualified table and not in the main table of the repository structure in Console).
    Minimum required field mapping: All non-qualifier fields of the qualified lookup table should be mapped. Qualifier field (Q) is optional.
    Now we are left with one field (Lookup [Price]) in destination which is still unmapped and we have no corresponding field left in the source to map it with. For this, go to source field tab and right click to create compound field 
    Lookup [Price] will be available in the source field and will get automatically mapped to the corresponding destination field.
    Note: If any one of the non-qualifier fields of the qualified lookup table is not yet mapped, we are not allowed to create the compound field.
    Do the Record Matching and Import the data into Data Manager Client by selecting the appropriate Import action.
    Final Qualified data in main table:
    The data is available in the Data Manager.
    P.S. If there is only one non-qualifier in a scenario, then the Compound field is not created and direct one to one mappings suffice.
    Please award me points if u find trhis blog as a value addition.
    Thanks & Regards ,
    Deepankar

  • One form for multiple tables

    Hey,
    I'm Pretty new to Access, as I've been given units to cover for my college course.  Unfortunately my tutor has had this dumped on him too (so he's kinda self teaching as he goes as well)....
    My issue is that I am trying to create a single form that will allow for all the information to be entered onto, which will then disseminate that info to the separate tables.
    The database is a simple film database that has 8 tables in total. Seven are data tables (Filminfo; Certificate; Genre; Series; Role; Producer; Actor) and the seventh is a central table that is filled with keys as a connection for them all. Series is also
    an off shoot from the filminfo table. I do have a pic, but I couldn't attach an image until my account was verified!
    This is how we were advised to set it out and as far as I can tell the other parts that have been built are working. I know there is a way to do it, but tonight the internet is failing me, that, or my lack of understanding
    exactly what I'm reading.
    Therefore any help that could be offered by anybody that can work out what I'm rambling on about, would be most appreciated (even more so if in layman's terms)

    What you have in relational database terms is a 7-way many-to-many relationship type between the 'referenced' tables.  This is what your eighth table is modelling by resolving the many-to-many relationship type into seven one-to-many relationship types.
    The problem here is that you are going to end up with a huge amount of 'redundancy' in the table modelling the relationship type.  At least some of the other tables represent entity types of which there can be more that one instance per film, e.g. there
    will be more than one actor in most films.  So let's say you record four actors for a film.  That means four rows in the table, but you then have to record the producer of the film four times as well, entering the same foreign key value into all
    four rows.  That's what's meant by redundancy.  Redundancy is a bad thing because it leaves the table open to the risk of 'update anomalies', e.g. two or more different producers could be inadvertently entered in different rows of the four. 
    Which is the correct one?  There's no way of knowing.
    So what relationship types do we really have here?  Let's concentrate on films and actors.  Each film can have many actors, and each actor can appear in many films (think Michael Caine!), so the relationship type is many-to-many.  Now a many-to-many
    relationship type cannot be modelled directly between two tables, it's done by a third table which resolves it into two one-to-many relationship types.  So in broad outline the tables would be like this:
    Films
    ....FilmID  (PK)
    ....Title
    ....ReleaseDate
    ....etc
    Actors
    ....ActorID  (PK)
    ....FirstName
    ....LastName
    ....etc
    and to model the relationship type:
    FilmActors
    ....FilmID  (FK)
    ....ActorID  (FK)
    ....Role
    The primary key of this last table is a composite one made up of the two foreign keys FilmID and ActorID.
    Karl's already given you a clue how to represent this; by means of a form/subform.  The form would be based on Films and the subform on FilmActors.  For a simple example of this basic type of many-to-many relationship type and how to represent it
    in a form/subform take a look at StuidentCourses.zip in my public databases folder at:
    https://onedrive.live.com/?cid=44CC60D7FEA42912&id=44CC60D7FEA42912!169
    If you have difficulty opening the link copy its text (NB, not the link location) and paste it into your browser's address bar.
    This little demo file illustrates a number of interfaces using forms, but just look at the conventional form/subform set-up.  The other two are just there because from time to time people have asked how to do it in these other ways.  Quite why they
    want to is another matter!
    My demo just has one subform, but there's no reason why you can't have multiple subforms in a films form of course, each for a different related entity type, each linked to the parent form on FilmID.  Tip:  to save space put ach subform on a separate
    page of a tab control, with the main film data in the parent form above the tab control so it's visible permanently as you move between different pages of the tab control.
    BTW in my StudentCourses demo try typing into the combo box in the subform the name of a new course not currently listed, French say, and see what happens.  Take a look at the code in the combo box's NotInList event procedure to see how it works. 
    For other uses of this event procedure take a look at the NotInlist demo file in the same OneDrive folder.
    I can sympathise with your tutor BTW.  My next door neighbour is head of the art department at a high school, but she has to teach maths also!
    PS:  I included Role as column in the FilmActors table, but what about Alec Guinness in Kind Hearts and Coronets? (google it if you're not familiar with the film)  How do you think you'd handle that situation?
    Ken Sheridan, Stafford, England

  • "Error in account determination:Table T030K key MWS"

    Hello Gurus,
    I would greatly appreciate some help on the following issues. The issue relates to tax on sales & purchases. The client charges a state tax for residents of certain states. In FTXP, I have used TAXUSJ and also tax code O1 (A/R Sales Tax, taxable). I have also put the state that charges the tax in the Tax Jurisdiction field in FTXP. In MW1 (Accrued: A/R Sales Tax 1) I have 6.500% for the state.
    On the SD side, the consultant is using the Acct Key of MWS for the Condition type UTX1 (hope that makes sense).
    When I try to release the billing document to accounting I get the following error:
    <b>Error in account determination:Table T030K key MWS</b>
    <b>Message no. FF709</b>
    <b>Diagnosis</b>
    In the chart of accounts to be posted to, no accounts are defined for the tax code you used.
    <b>Procedure</b>
    Contact your system administrator. Define the accounts to which a tax posting is to be made with the tax code entered in Customizing for taxes on sales/purchases. To do this, choose Maintain entries (F5).
    I went to OBYZ, double clicked on PROCEDURES, then selected TAXUSJ and double clicked on CONTROL. Once inside, I then <u>changed the account key for condition type JR1 from MW1 to MWS  and saved it.</u>
    But when I went back to release the billing document to accounting, I now get a new error:
    <b>System error in routine FI_TAX_GET_TXJCD_LEVELS error code 2 function builder TAX2</b>Message no. FF769
    <b>Diagnosis</b>
    A serious program error has occurred.
    <b>System Response</b>
    The current processing step cannot be carried out.
    <b>Procedure</b>
    If the program error occurred
    - In a standard SAP program, then create an OSS message
    - In a program you wrote yourself, then correct the program.
    I went back and changed MWS back to MW1 but still get this new error. Does anyone know what could be wrong?
    Thanks in advance!

    Hi United99
    To tell u in layman words, u r missing in Account determinatin. Tax determination is almost same as Pricing Procidur in SD for which we assign GL accounts to transaction key. As there are ERL, ERF, keys u have MWS (Output Tax)key assigned to Tax procidure TAXUS (check this at defining tax procidre.
    Now we assign GL accounts to this Transaction key MWS as OB40. U assign GL account at OB40. Your issue will be resolved. But..
    MWS is default Key, i suggest not to distrub it but copy it and assign it to your procidure and to this key assig GL account at OB40.
    Please assign points if useful
    Satish

  • Implicit fact and Alias table with

    Hi Gurus,
    May i know what is the use of alias table and implicit fact with some real time scenarios for better under standing.
    Thanks,
    Rafi

    For alias tables, refer to the oracle's documentation here which is explained in great detail: ( http://docs.oracle.com/cd/E14571_01/bi.1111/e10540/physicallayer.htm )
    For implicit fact refer to the oracle's documentation here: ( http://oracle-bi.siebelunleashed.com/articles/implicit-fact-column/ ) . To define in layman's terms, implicit fact is set on a subject area when you are reporting out of dimensions in the report without any fact columns and you want the fact table to be part of the report's query.

  • How to Join 2 tables if datetime diff is within range?

    Hi, problem figuring out how to join tables if the datediff ? is within range.  
    There are 3 tables  a main table then 2 sub tables, these are measurements with a date/time value.  Im trying to take the datetime entries from the 2 sub tables and match them, then add  them to the main table
    subtable1 has SampleID, datetime, 
    subtable2 has TesterID, datetime
    maintable has SampleID, datetime, TesterID, datetime
    so the maintable would always have the sampleID value and datetime regardless.  but it would only JOIN the testerID and its datetime, only IF these 2 datetime values were close, and the datediff interval was set to 1 second.
     its a function call that fills the main tables matching columns but only if the testerIDs datetime is close to the sampleIDs time????
    thanks

    Please post DDL, so that people do not have to guess what the keys, constraints, Declarative Referential Integrity, data types, etc. in your schema are. Learn how to follow ISO-11179 data element naming conventions and formatting rules. Temporal data should
    use ISO-8601 formats. Code should be in Standard SQL as much as possible and not local dialect. 
    This is minimal polite behavior on SQL forums. Thanks for making us do your typing and DDL for you :( 
    >> There are 3 tables; a main table then 2 sub tables, <<
    Really? Main table? Sub-tables? There are no such terms in SQL. Trust me on that; I helped with the Standards :) We have referenced referencing tables in the DRI model of SQL. 
    >> these are measurements with a date/time value. << 
    ++Both of them, the measurementID and otherID  have a millisecond timestamp, the goal is to associate these 2 measurements, they are coming from different systems.  
    Do you really need the full timestamp down to nanoseconds and not just a DATE? Oh, we also have no sample data :( So we have to guess at precision, constraints, keys, etc. 
    See why we need DDL? Is this relationship 1:1, 1:M OR N:1? My guess is that a tester can do many samples, but a sample has only one tester. And what were those two timestamps? Not redundant copies from the referenced tables!! That would non-normalized. 
    >> so the Sample_Tester_Relationship would always have the sample_nbr value and <something>_timestamp regardless. But it would only JOIN the tester_id and its DATETIME, only IF these 2 DATETIME values were close, and the DATEDIFF interval was set
    to 1 second. <<
    No, you do not do a JOIN inside a base table. That is VIEW.
    >> It is a function [sic: procedure, Functions return scalar values] call that fills the main tables matching columns but only if the tester_id DATETIME is close to the sample_id time? <<
    No sample data, no sample code and no business rules about ties. Here is my final guess at normalizing this problem: 
    CREATE TABLE Samples
    (sample_nbr CHAR(12) NOT NULL PRIMARY KEY,
    CREATE TABLE Testers
    (tester_id CHAR(10) NOT NULL PRIMARY KEY,
    CREATE TABLE Sample_Tester_Relationship --- needs a real name!
    (sample_nbr CHAR(12) NOT NULL PRIMARY KEY
     REFERENCES Samples
      ON UPDATE CASCADE
      ON DELETE CASCADE,
     tester_id CHAR(10) NOT NULL
     REFERENCES Testers
      ON UPDATE CASCADE
      ON DELETE CASCADE,
     PRIMARY KEY (sample_nbr, tester_id),
     tester_timestamp DATETIME2(1) DEFAULT CURRENT_TIMESTAMP NOT NULL,
     sample_timestamp DATETIME2(1) DEFAULT CURRENT_TIMESTAMP NOT NULL,
     CHECK (ABS (DATEDIFF (SECOND, tester_timestamp, sample_timestamp) <= 1) );
    Most of the work in SQL is in DDL, not DML. 
    I will try this out and advise.
    --CELKO-- Books in Celko Series for Morgan-Kaufmann Publishing: Analytics and OLAP in SQL / Data and Databases: Concepts in Practice Data / Measurements and Standards in SQL SQL for Smarties / SQL Programming Style / SQL Puzzles and Answers / Thinking in
    Sets / Trees and Hierarchies in SQL
    Dear Celko
    thank you for this constructive feedback, clearly you're an expert and I am going to be better for following this advice.  In the 1970s I was told to avoid going into software because Japan had invented a way for computers to write their own code, so
    programmers would be out of a job and the field would be a dead-end.  I didnt formally begin training in sw until the 90s.  SQL is one of those last frontiers i never trained in.  but now Im trying to learn and find anything i can read to absorb
    it.
    I posted in laymans words because i thought thats how you experts would prefer it described, clearly not.  I will now take the time to read each reference to the standards...

  • MB5B Report table for Open and Closing stock on date wise

    Hi Frds,
    I am trying get values of Open and Closing stock on date wise form the Table MARD and MBEW -Material Valuation but it does not match with MB5B reports,
    Could anyone suggest correct table to fetch the values Open and Closing stock on date wise for MB5B reports.
    Thanks
    Mohan M

    Hi,
    Please check the below links...
    Query for Opening And  Closing Stock
    Inventory Opening and Closing Stock
    open stock and closing stock
    Kuber

  • Error while dropping a table

    Hi All,
    i got an error while dropping a table which is
    ORA-00600: internal error code, arguments: [kghstack_free1], [kntgmvm: collst], [], [], [], [], [], [], [], [], [], []
    i know learnt that -600 error is related to dba. now how to proceed.
    thanks and regards,
    sri ram.

    00600 errors should be raised as service request with Oracle as it implies some internal bug.
    You can search oracle support first to see if anyone has had the same class of 00600 error, and then if not (and therefore no patch) raise your issue with Oracle.
    http://support.oracle.com

  • Logical level in Fact tables - best practice

    Hi all,
    I am currently working on a complex OBIEE project/solution where I am going straight to the production tables, so the fact (and dimension) tables are pretty complex since I am using more sources in the logical tables to increase performance. Anyway, what I am many times struggling with is the Logical Levels (in Content tab) where the level of each dimension is to be set. In a star schema (one-to-many) this is pretty straight forward and easy to set up, but when the Business Model (and physical model) gets more complex I sometimes struggle with the aggregates - to get them work/appear with different dimensions. (Using the menu "More" - "Get levels" does not allways give the best solution......far from). I have some combinations of left- and right outer join as well, making it even more complicated for the BI server.
    For instance - I have about 10-12 different dimensions - should all of them allways be connected to each fact table? Either on Detail or Total level. I can see the use of the logical levels when using aggregate fact tables (on quarter, month etc.), but is it better just to skip the logical level setup when no aggregate tables are used? Sometimes it seems like that is the easiest approach...
    Does anyone have a best practice concerning this issue? I have googled for this but I haven't found anything good yet. Any ideas/articles are highly appreciated.

    Hi User,
    For instance - I have about 10-12 different dimensions - should all of them always be connected to each fact table? Either on Detail or Total level.It not necessary to connect to all dimensions completely based on the report that you are creating ,but as a best practice we should maintain all at Detail level only,when you are mentioning any join conditions in physical layer
    for example for the sales table if u want to report at ProductDimension.ProductnameLevel then u should use detail level else total level(at Product,employee level)
    Get Levels. (Available only for fact tables) Changes aggregation content. If joins do not exist between fact table sources and dimension table sources (for example, if the same physical table is in both sources), the aggregation content determined by the administration tool will not include the aggregation content of this dimension.
    Source admin guide(get level definition)
    thanks,
    Saichand.v

  • Rendering xml-table into logical filename in SAP R/3

    Hi,
    I am trying to translate an xml-table with bytes into a logical filepath in SAP R3.
    Do I have to use the method gui-download or shall I loop the internal xml-table?
    When I tried to loop the xml-table into a structure, and then transfering the structure into the logical filename, I get problems with the line breaks in my xml-file. How do I get the lines to break exactly the same as I wrote them in my ABAP-code?
    Edited by: Kristina Hellberg on Jan 10, 2008 4:24 PM

    I believe you posted in the wrong forum.
    This forum is dedicated to development and deployment of .Net applications that connect and interact with BusinessObjects Enterprise, BusinessObjects Edge, or Crystal Reports Server. This includes the development of applications using the BusinessObjects Enterprise, Report Application Server, Report Engine, and Web Services SDKs.
    Ludek

  • Can you check for data in one table or another but not both in one query?

    I have a situation where I need to link two tables together but the data may be in another (archive) table or different records are in both but I want the latest record from either table:
    ACCOUNT
    AccountID     Name   
    123               John Doe
    124               Jane Donaldson           
    125               Harold Douglas    
    MARKETER_ACCOUNT
    Key     AccountID     Marketer    StartDate     EndDate
    1001     123               10526          8/3/2008     9/27/2009
    1017     123               10987          9/28/2009     12/31/4712    (high date ~ which means currently with this marketer)
    1023     124               10541          12/03/2010     12/31/4712
    ARCHIVE
    Key     AccountID     Marketer    StartDate     EndDate
    1015     124               10526          8/3/2008     12/02/2010
    1033     125               10987         01/01/2011     01/31/2012  
    So my query needs to return the following:
    123     John Doe                        10526     8/3/2008     9/27/2009
    124     Jane Donaldson             10541     12/03/2010     12/31/4712     (this is the later of the two records for this account between archive and marketer_account tables)
    125     Harold Douglas               10987          01/01/2011     01/31/2012     (he is only in archive, so get this record)
    I'm unsure how to proceed in one query.  Note that I am reading in possibly multiple accounts at a time and returning a collection back to .net
    open CURSOR_ACCT
              select AccountID
              from
                     ACCOUNT A,
                     MARKETER_ACCOUNT M,
                     ARCHIVE R
               where A.AccountID = nvl((select max(M.EndDate) from Marketer_account M2
                                                    where M2.AccountID = A.AccountID),
                                                      (select max(R.EndDate) from Archive R2
                                                    where R2.AccountID = A.AccountID)
                   and upper(A.Name) like parameter || '%'
    <can you do a NVL like this?   probably not...   I want to be able to get the MAX record for that account off the MarketerACcount table OR the max record for that account off the Archive table, but not both>
    (parameter could be "DO", so I return all names starting with DO...)

    if I understand your description I would assume that for John Dow we would expect the second row from marketer_account  ("high date ~ which means currently with this marketer"). Here is a solution with analytic functions:
    drop table account;
    drop table marketer_account;
    drop table marketer_account_archive;
    create table account (
        id number
      , name varchar2(20)
    insert into account values (123, 'John Doe');
    insert into account values (124, 'Jane Donaldson');
    insert into account values (125, 'Harold Douglas');
    create table marketer_account (
        key number
      , AccountId number
      , MktKey number
      , FromDt date
      , ToDate date
    insert into marketer_account values (1001, 123, 10526, to_date('03.08.2008', 'dd.mm.yyyy'), to_date('27.09.2009', 'dd.mm.yyyy'));
    insert into marketer_account values (1017, 123, 10987, to_date('28.09.2009', 'dd.mm.yyyy'), to_date('31.12.4712', 'dd.mm.yyyy'));
    insert into marketer_account values (1023, 124, 10541, to_date('03.12.2010', 'dd.mm.yyyy'), to_date('31.12.4712', 'dd.mm.yyyy'));
    create table marketer_account_archive (
        key number
      , AccountId number
      , MktKey number
      , FromDt date
      , ToDate date
    insert into marketer_account_archive values (1015, 124, 10526, to_date('03.08.2008', 'dd.mm.yyyy'), to_date('02.12.2010', 'dd.mm.yyyy'));
    insert into marketer_account_archive values (1033, 125, 10987, to_date('01.01.2011', 'dd.mm.yyyy'), to_date('31.01.2012', 'dd.mm.yyyy'));
    select key, AccountId, MktKey, FromDt, ToDate
         , max(FromDt) over(partition by AccountId) max_FromDt
      from marketer_account
    union all
    select key, AccountId, MktKey, FromDt, ToDate
         , max(FromDt) over(partition by AccountId) max_FromDt
      from marketer_account_archive;
    with
    basedata as (
    select key, AccountId, MktKey, FromDt, ToDate
      from marketer_account
    union all
    select key, AccountId, MktKey, FromDt, ToDate
      from marketer_account_archive
    basedata_with_max_intervals as (
    select key, AccountId, MktKey, FromDt, ToDate
         , row_number() over(partition by AccountId order by FromDt desc) FromDt_Rank
      from basedata
    filtered_basedata as (
    select key, AccountId, MktKey, FromDt, ToDate from basedata_with_max_intervals where FromDt_Rank = 1
    select a.id
         , a.name
         , b.MktKey
         , b.FromDt
         , b.ToDate
      from account a
      join filtered_basedata b
        on (a.id = b.AccountId)
    ID NAME                     MKTKEY FROMDT     TODATE
    123 John Doe                  10987 28.09.2009 31.12.4712
    124 Jane Donaldson            10541 03.12.2010 31.12.4712
    125 Harold Douglas            10987 01.01.2011 31.01.2012
    If your tables are big it could be necessary to do the filtering (according to your condition) in an early step (the first CTE).
    Regards
    Martin

  • Can not insert/update data from table which is created from view

    Hi all
    I'm using Oracle database 11g
    I've created table from view as the following command:
    Create table table_new as select * from View_Old
    I can insert/update data into table_new by command line.
    But I can not Insert/update data of table_new by SI Oject Browser tool or Oracle SQL Developer tool .(read only)
    Anybody tell me, what's happend? cause?
    Thankyou
    thiensu
    Edited by: user8248216 on May 5, 2011 8:54 PM
    Edited by: user8248216 on May 5, 2011 8:55 PM

    I can insert/update data into table_new by command line.
    But I can not Insert/update data of table_new by SI Oject Browser tool or Oracle SQL Developer tool .(read only)so what is wrong with the GUI tools & why posting to DATABASE forum when that works OK?

  • Not able to refresh the data in a table

    Hi In my application i fill data in a table on clikc of a button ..
    Following are the line of code i have user
    RichColumn richcol = (RichColumn)userTableData.getChildren().get(0);                                           (fetch the first column of my table)
    richcol.getChildren().add(reportedByLabel);                                                                   (reportedByLabel is a object of RichInputText)
    AdfFacesContext adfContext1 = AdfFacesContext.getCurrentInstance();
    adfContext1.addPartialTarget(richcol);
    adfContext1.addPartialTarget(userTableData);
    But on submit of that button table data is not refreshed after adding partial trigger on that table as well as that column also .... any idea??
    Edited by: Shubhangi m on Jan 27, 2011 3:50 AM

    Hi,
    The Code that you have shown adds an additional inputText component to the first column of a table.
    Is that your intention?
    If yes, please use the following code snippet to achieve your functionality:
    JSPX Code:
    <af:form id="f1">
    <af:commandButton text="Add Column" id="cb1"
    actionListener="#{EmployeesTableBean.onAddColumn}"/>
    <af:table value="#{bindings.Employees.collectionModel}" var="row"
    rows="#{bindings.Employees.rangeSize}"
    emptyText="#{bindings.Employees.viewable ? 'No data to display.' : 'Access Denied.'}"
    fetchSize="#{bindings.Employees.rangeSize}"
    rowBandingInterval="0"
    selectedRowKeys="#{bindings.Employees.collectionModel.selectedRow}"
    selectionListener="#{bindings.Employees.collectionModel.makeCurrent}"
    rowSelection="single" id="t1"
    binding="#{EmployeesTableBean.table}">
    <af:column sortProperty="EmployeeId" sortable="false"
    headerText="#{bindings.Employees.hints.EmployeeId.label}"
    id="c1">
    <af:outputText value="#{row.EmployeeId}" id="ot2">
    <af:convertNumber groupingUsed="false"
    pattern="#{bindings.Employees.hints.EmployeeId.format}"/>
    </af:outputText>
    </af:column>
    <af:column sortProperty="FirstName" sortable="false"
    headerText="#{bindings.Employees.hints.FirstName.label}"
    id="c2">
    <af:outputText value="#{row.FirstName}" id="ot1"/>
    </af:column>
    <af:column sortProperty="LastName" sortable="false"
    headerText="#{bindings.Employees.hints.LastName.label}"
    id="c3">
    <af:outputText value="#{row.LastName}" id="ot3"/>
    </af:column>
    </af:table>
    </af:form>
    Bean:
    public class EmployeesTableBean {
    private RichTable table;
    public EmployeesTableBean() {
    public void setTable(RichTable table) {
    this.table = table;
    public RichTable getTable() {
    return table;
    public void onAddColumn(ActionEvent actionEvent) {
    RichInputText newRichInputText = new RichInputText();
    newRichInputText.setId("new");
    newRichInputText.setValue("Name:");
    RichColumn richcol = (RichColumn)table.getChildren().get(0);
    richcol.getChildren().add(newRichInputText);
    AdfFacesContext adfContext1 = AdfFacesContext.getCurrentInstance();
    adfContext1.addPartialTarget(table);
    Thanks,
    Navaneeth

  • Unable to capture the adf table column sort icons using open script tool

    Hi All,
    I am new to OATS and I am trying to create script for testing ADF application using open script tool. I face issues in recording two events.
    1. I am unable to record the event of clicking adf table column sort icons that exist on the column header. I tried to use the capture tool, but that couldn't help me.
    2. The second issue is I am unable to capture the panel header text. The component can be identified but I was not able to identify the supporting attribute for the header text.

    Hi keerthi,
    1. I have pasted the code for the first issue
    web
                             .button(
                                       122,
                                       "/web:window[@index='0' or @title='Manage Network Targets - Oracle Communications Order and Service Management - Order and Service Management']/web:document[@index='0' or @name='1824fhkchs_6']/web:form[@id='pt1:_UISform1' or @name='pt1:_UISform1' or @index='0']/web:button[@id='pt1:MA:0:n1:1:pt1:qryId1::search' or @value='Search' or @index='3']")
                             .click();
                        adf
                        .table(
                                  "/web:window[@index='0' or @title='Manage Network Targets - Oracle Communications Order and Service Management - Order and Service Management']/web:document[@index='0' or @name='1c9nk1ryzv_6']/web:ADFTable[@absoluteLocator='pt1:MA:n1:pt1:pnlcltn:resId1']")
                        .columnSort("Ascending", "Name" );
         }

  • How to Activate the Table which Doesn't Exist - Can I create Table in SE11

    Hi Gurus,
    I am a BASIS Person ..so, I don't have much idea about ABAP Developement side . But, while applying patches to one of  the SAP Component, I get the following error message ( usr\sap\trans\log ) :
    Phase Import_Proper
    >>>
    SAPAIBIIP7.BID
    1 ED0301 *************************************************************************
    1 EDO578XFollowing objects not activated/deleted or activated/deleted w. warning:
    1EEDO519X"Table" "WRMA_S_RESULT_DATA" could not be activated
    1EEDO519 "Table Type" "WRMA_TT_UPDTAB_DSO" could not be activated
    1 ED0313 (E- Row type WRMA_S_RESULT_DATA is not active or does not exist )
    2WEDO517 "Domain" "WRMA_KAPPL" was activated (error in the dependencies)
    2WEDO517 "Data Element" "/RTF/DE_FISCPER" was activated (error in the dependencies)
    2WEDO517 "Data Element" "/RTF/DE_FISCVARNT" was activated (error in the dependencies)
    2WEDO517 "Data Element" "WRMA_DE_RCLASV" was activated (error in the dependencies)
    2WEDO517 "Data Element" "WRMA_DE_RMA_AMNT" was activated (error in the dependencies)
    2WEDO517 "Data Element" "WRMA_DE_RMA_INV" was activated (error in the dependencies)
    2WEDO517 "Data Element" "WRMA_DE_RMA_OBJ" was activated (error in the dependencies)
    2WEDO549 "Table" "V_WRMA_TRCLASVER" was activated (warning for the dependent tables)
    2WEDO549 "Table" "V_WRMA_TRCLASVPE" was activated (warning for the dependent tables)
    2WEDO549 "Table" "V_WRMA_TRCLPERCL" was activated (warning for the dependent tables)
    2WEDO549 "Table" "WRMA_S_STOCK_SELECTION_DATE" was activated (warning for the dependent tables)
    <<<
    Now, I checked Domain WRMA_KAPPL  in SE11 and it diplays that it is partially active :
    " The obeject is marked partially active for the following reasons :
       Errors occured when adjusting dependent Objects to a change. Some of the dependent objects could not
       be adjusted to this change. To adjust the dependent objects, you must perform the following actions
       the next time you activate this object.
                        - SET THE SYNP TIME STAMP IN THE RUNTIME OBJECT OFALL DEPENDENT OBJECTS
    My question is - how to activate this SYNP Time Stamp.. ?
                          - Once I activate this SYNP Time Stamp.. How do I activate this ..in SE11 or some other
                            TCode..
    Any / ALL Help is highy appreciated and would be rewarded with appropriate Reward Points.
    Thanks,
    Regards,
    - Ishan
    >> Ok, Now I forcefully activated the Domain and then all the 5  Data Elements in SE11 ...now I have error
         that table  WRMA_S_RESULT_DATA  and WRMA_TT_UPDTAB_DSO Could not be activated...
         ..And in SE11 > Input the first / Second Table  > Excute > Output : Table Does not Exist !!!
          - Now, How do I activate something which does not exist in the first place ?
        Any Input about this ?
    Thanks,
    Regards,
    - Ishan
    Edited by: ISHAN P on Jun 19, 2008 3:44 PM

    Read note 1152612 - Incorrect component type in structure WRMA_S_RESULT_DATA.
    The error occurs in Release BI_CONT 703 Support Package 9.
    If you require an advance correction for Support Package 9, make the following manual changes if the InfoObject 0TCTLSTCHG is inactive in your system.
    1. Use transaction RSA1 to call the Data Warehousing Workbench.
    2. Check whether the InfoObject 0TCTLSTCHG was already activated by the compilation process the first time you called transaction RSA1.  You can use transaction RSD1 to check this by entering 0TCTLSTCHG in the "InfoObject" field and choosing "Display".  Proceed as follows if the InfoObject is still not active:
    3. Go to BI Content activation.
    4. Select "Object Types".
    5. Open the "InfoObjects" tree and select "Objects".
    6. Search for the InfoObject 0TCTLSTCHG and select it.
    7. Transfer the object and activate it.
    These steps activate the required component type /BI0/OITCTLSTCHG. You can check again in transaction RSD1 to ensure that the InfoObject 0TCTLSTCHG is available as active.
    Jacek Fornalczyk

Maybe you are looking for

  • Can't Get Back to My Mac

    I have tried without success to get Back to My Mac working since downgrading from MobileMe to iCloud.  It can find some of the Macs on my network but not others, and it's never the same ones that get found or not found from one day to the next.  In a

  • Document type WE.

    Hi, When we do the Good Receipt then accounting document will be created using document type WE. Do we have to do some configuration according to this? Thanks.

  • Graph range problem with tab control

    Hello. I am using LabVIEW 8.2.1 I have five XY-graphs in a tab control, with five tabs. One graph in each tab. When running, theese graphs is constantly written to, but the user is only allowed to watch one at a time. The program also contain a maxim

  • Help needed please for a newbie

    hi all i recently recoded some fottage on my mobile phone and the resulting file is a mp4 now when i put the mp4 into adobe it only recognises it as a audio file i want to be able to use the file in adobe for editing and the like so how do i get adob

  • Service level attribute

    Hi gurus, i want to display the attribute in my query. but when right click the char and select the attribute, ( it is in the service level - unit of measure attribute) not giving the values. it is giving the 0.00 values. but in the master maintenanc