GL_INTERFACE_CONTROL TABLE 설명

제품 : FIN_GL
작성날짜 : 2003-11-18
GL_INTERFACE_CONTROL TABLE 설명
=============================
PURPOSE
GL_INTERFACE_CONTROL table에 대해 알아보도록 한다.
Explanation
gl_interface_control table은 gl_interface 까지 data가 넘어 온 후 journal import program 수행 시 에 data 가 insert 되어진다.
즉 interface 만 완료 후 journal import 를 수행하지 않았을 경우에는 아무런 data도 생기지 않는다.
(ap의 경우 transfer parameter option에서 'Submit Journal Import'에 check 하지 않았을 경우에는 data 가 생기지 않는다.)
Status Column 은 다음과 같은 code를 가진다.
Status Meaning
P Indicates the source is available to be used for an Import.
Sources that have a status 'P' are not currently being imported
and have not been selected for processing by a pending import.
S Indicates the source has been selected for processing by
Journal Import.
I Indicates the source is currently being processed by
Journal Import.
Example
Reference Documents
Note 1060932.6

Can you try to use to_char for Interface run id as shown below
ln_reg_id := fnd_request.submit_request
(application => 'SQLGL', -- application short name
program => 'GLLEZL', -- program short name
description => NULL, -- program name
start_time => NULL, -- start date
sub_request => FALSE, -- sub-request
argument1 =>to_char( l_iface_run_id), -- interface run id
argument2 => fnd_profile.VALUE ('GL_SET_OF_BKS_ID'), -- set of books id
argument3 => 'N', -- error to suspense flag
argument4 => NULL, -- from accounting date
argument5 => NULL, -- to accounting date
argument6 => 'N', -- create summary flag
argument7 => 'N' -- import desc flex flag
);

Similar Messages

  • How to drop a table in another schema.

    Hi there, was looking to drop a table that is currently created under GL schema when consolidation process is done. These tables are kind of temporary and stay there if consolidation process doesn't complete successfully.
    So, I'm try to execute a script that would drop the table but the particularity is that this drop isn'T executed by apps but with bolinf (custom user).
    So, while trying to delete it with bolinf, I always get an error message that the table doesn't exists or so.
    Here's the code
    r_gl_int_ctrl.INTERFACE_TABLE_NAME is retreived from gl_interface_control table and from INTERFACE_TABLE_NAME column.
    EXECUTE IMMEDIATE 'DROP TABLE GL.' || r_gl_int_ctrl.INTERFACE_TABLE_NAME;
    and this generates the following error : ORA-00942: table or view does not exist
    R12.0.6
    RDBMS 10.2.0.3.0
    Custom package under bolinf schema is trying to drop GL.GL_CONS_INTERFACE_38190 and it's not working. Nevertheless, if I install this package under apps, table get dropped.
    Is there a way that I can drop a GL table like GL.GL_CONS_INTERFACE_38190 from BOLINF schema ? If so, what privileges/grant should user BOLINF have in order to drop this table from a GL Schema ?
    Thank you.

    Hi;
    You need to give drop grant to your user.Easy way is to give
    GRANT DROP ANY TABLE TO USER; << but its mean it will give drop grant for all table.
    Also see:
    OERR: ORA 942 "table or view does not exist" [ID 18536.1]
    Regard
    Helios

  • AP to GL Transfer: How to have a custom Journal Header name in GL.

    Hi,
    Does any one know how to have a custom Journal Header name while transferring the Journals from AP to GL?
    Assume that I am transferring the Journals which belongs to a Single Invoice from AP to GL. When these Journals are transferred from AP to GL, Oracle will create a Journal Header with the default naming convention. The Oracle generated Header name will be as shown below.
    *123456 Purchase Invoices USD*. Where 123456 is AE_HEADER_ID and USD is the Invoice Currency code.
    Our requirement is, we want the Journal Header name something like *"Hello World"+TIMESTAMP*, where TIMESTAMP is the date when the user has ran the AP to GL transfer program.
    I know that when building the AAD(Application Accounting Definition) we can define the Journal Header descriptions and Journal Lines descriptions. But, I did not see any setup option to have the custom Journal Header name in GL.
    Any suggestions on this issue are highly appreciated.
    Advanced Thanks,
    Lokesh.

    Hi Lokesh,
    The best way to have a custom journal header name is using hook package.
    When data move from XLA to GL, every time it creates a new dynamic table as GL_INTERFACE table but the name is different (XLA_GLT_XXXXX).
    This table information stores in 'gl_interface_control' table but after use the record has deleted from 'gl_interface_control'.
    So we can't see table information after finishing create accounting program.
    What the modification you will made in this table it will affect GL base table. So to made custom journal header name, we need to change Reference4 column of dynamic table.
    Here is the code to get the table name and through procedure you can update the Reference4 column.
    FUNCTION pre_module_hook(run_id IN NUMBER,
                   errbuf IN OUT NOCOPY VARCHAR2) RETURN BOOLEAN IS
    chr_errbuf varchar2(50);
    chr_retcode varchar2(50);
    l_interface_Table_name varchar2(2000);
    BEGIN
         SELECT interface_table_name
    into l_interface_Table_name
    FROM gl_interface_control
    WHERE interface_run_id = run_id;
         BEGIN     
    xxdb_gl_icaccntgdist_pkg.gl_int_jrnl_update (chr_errbuf,chr_retcode,l_interface_Table_name);
    RETURN(TRUE);
         END;
    END pre_module_hook;
    Make these changes in GL_IMPORT_HOOK_PKG.pre_module_hook function.
    You can make the changes through calling procedure using dynamic query under the function pre_module_hook.
    Thanks,
    Gaurav K
    Edited by: 972729 on Nov 22, 2012 4:46 AM

  • Journal import fails when called from PLSQL

    Hi,
    When journal import is called from plsql code it is failing with error in 'gllacc() Function returning without value and no data found'.
    Same transaction is run succesfully from front end.
    I checked both gl_interface and gl_interface_control table but couldnt find the issue.
    Any info on this would be very helpful.
    Thanks
    Sandhya

    FOR l_rec IN (SELECT ledger_id,group_id from apps.gl_interface
    WHERE status='NEW'
    AND user_je_source_name='GIS_DATA_CONVERSION'
    GROUP BY ledger_id,group_id
    ORDER BY group_id
    LOOP
    apps.gl_journal_import_pkg.populate_interface_control (user_je_source_name => 'GIS_DATA_CONVERSION',
    GROUP_ID => l_rec.group_id,
    set_of_books_id => l_rec.ledger_id,
    interface_run_id =>vl_interface_id,
    table_name => 'GL_INTERFACE',
    processed_data_action=>'D'
    COMMIT;
    vl_request_id := apps.fnd_request.submit_request (application => 'SQLGL', -- application short name
    program => 'GLLEZLSRS', -- program short name
    description => NULL, -- program name
    start_time => NULL, -- start date
    sub_request => FALSE, -- sub-request
    argument1 => 2065, --Data access set id
    argument2 => 'GIS_DATA_CONVERSION', --Source
    argument3 => l_rec.ledger_id, -- set of books id
    argument4 => l_rec.group_id,
    argument5 => 'N', -- error to suspense flag
    argument6 => NULL, -- create summary flag
    argument7 => 'N' -- import desc flex flag
    COMMIT;
    IF ( vl_request_id = 0 ) THEN
    xxgis.gis_conv_util_pkg.debug_print_p(1,'FND_LOG','E001: Journal Import Submission Failed. ' || SQLERRM);
    retcode := 2;
    EXIT;
    ELSE
    xxgis.gis_conv_util_pkg.debug_print_p(1,'FND_LOG','P001: Submitted Journal Import Program for group id: ' || l_rec.group_id ||
    'and ledger :'||l_rec.ledger_id|| ', Request ID: ' || vl_request_id);
    END IF;
    END LOOP;

  • GL Balances Conversion

    Hello,
    I am working on an R12 implementation project. The legacy system is Oracle apps(11.5.x). I have to work on GL Balances converison. we want to get all the journal details from march 2009. i dont understand do we also get the balances since then??? what happens to the balances before march 2009?? I am new to this process. Any help is appreciated.
    Can i get the details of
    1. interface table
    2. import program
    Bhavani

    Hi, you'll have to use GL_INTERFACE and/or GL_INTERFACE_CONTROL table to import journal entries into general ledger.
    Also, Journal Import program will be used to import data from gl_interface and move then into gl_je_batches, gl_je_headers and gl_je_lines.
    you'll then decide if you post journal entries automatically or not
    So,
    1 - Import data into gl_interface and gl_interface_control
    2 - Run journal import process
    3 - Run the AutoPosting program to post Batches that were created in step 2 and are in Unposted Status.
    GL Interface table doc can be found in Oracle General Ledger User guide or in MyOracleSupport web site.

  • Multi Table GL Import

    We are trying to use a custom table for importing/posting journals in GL. Is it sufficient to enter the custom table name in GL_INTERFACE_CONTROL? Or any other changes necessary?
    Thanks
    Raghu.

    Hi,
    use internal table to hold the data.
    Pass this internal table to a memory ID and export it.
    Then import the same at the other end.
    Check the sample code.
    DATA : BEGIN OF ITABY OCCURS 0,
    END OF ITABY.
    EXPORT ITABY TO MEMORY ID 'ABC'.
    at other end use
    DATA : BEGIN OF ITABY OCCURS 0,
    END OF ITABY.
    IMPORT ITABY FROM MEMORY ID 'ABC.'
    Let me know if you have any queries.
    Regards,
    Raj

  • MB5B Report table for Open and Closing stock on date wise

    Hi Frds,
    I am trying get values of Open and Closing stock on date wise form the Table MARD and MBEW -Material Valuation but it does not match with MB5B reports,
    Could anyone suggest correct table to fetch the values Open and Closing stock on date wise for MB5B reports.
    Thanks
    Mohan M

    Hi,
    Please check the below links...
    Query for Opening And  Closing Stock
    Inventory Opening and Closing Stock
    open stock and closing stock
    Kuber

  • Error while dropping a table

    Hi All,
    i got an error while dropping a table which is
    ORA-00600: internal error code, arguments: [kghstack_free1], [kntgmvm: collst], [], [], [], [], [], [], [], [], [], []
    i know learnt that -600 error is related to dba. now how to proceed.
    thanks and regards,
    sri ram.

    00600 errors should be raised as service request with Oracle as it implies some internal bug.
    You can search oracle support first to see if anyone has had the same class of 00600 error, and then if not (and therefore no patch) raise your issue with Oracle.
    http://support.oracle.com

  • Logical level in Fact tables - best practice

    Hi all,
    I am currently working on a complex OBIEE project/solution where I am going straight to the production tables, so the fact (and dimension) tables are pretty complex since I am using more sources in the logical tables to increase performance. Anyway, what I am many times struggling with is the Logical Levels (in Content tab) where the level of each dimension is to be set. In a star schema (one-to-many) this is pretty straight forward and easy to set up, but when the Business Model (and physical model) gets more complex I sometimes struggle with the aggregates - to get them work/appear with different dimensions. (Using the menu "More" - "Get levels" does not allways give the best solution......far from). I have some combinations of left- and right outer join as well, making it even more complicated for the BI server.
    For instance - I have about 10-12 different dimensions - should all of them allways be connected to each fact table? Either on Detail or Total level. I can see the use of the logical levels when using aggregate fact tables (on quarter, month etc.), but is it better just to skip the logical level setup when no aggregate tables are used? Sometimes it seems like that is the easiest approach...
    Does anyone have a best practice concerning this issue? I have googled for this but I haven't found anything good yet. Any ideas/articles are highly appreciated.

    Hi User,
    For instance - I have about 10-12 different dimensions - should all of them always be connected to each fact table? Either on Detail or Total level.It not necessary to connect to all dimensions completely based on the report that you are creating ,but as a best practice we should maintain all at Detail level only,when you are mentioning any join conditions in physical layer
    for example for the sales table if u want to report at ProductDimension.ProductnameLevel then u should use detail level else total level(at Product,employee level)
    Get Levels. (Available only for fact tables) Changes aggregation content. If joins do not exist between fact table sources and dimension table sources (for example, if the same physical table is in both sources), the aggregation content determined by the administration tool will not include the aggregation content of this dimension.
    Source admin guide(get level definition)
    thanks,
    Saichand.v

  • Rendering xml-table into logical filename in SAP R/3

    Hi,
    I am trying to translate an xml-table with bytes into a logical filepath in SAP R3.
    Do I have to use the method gui-download or shall I loop the internal xml-table?
    When I tried to loop the xml-table into a structure, and then transfering the structure into the logical filename, I get problems with the line breaks in my xml-file. How do I get the lines to break exactly the same as I wrote them in my ABAP-code?
    Edited by: Kristina Hellberg on Jan 10, 2008 4:24 PM

    I believe you posted in the wrong forum.
    This forum is dedicated to development and deployment of .Net applications that connect and interact with BusinessObjects Enterprise, BusinessObjects Edge, or Crystal Reports Server. This includes the development of applications using the BusinessObjects Enterprise, Report Application Server, Report Engine, and Web Services SDKs.
    Ludek

  • Can you check for data in one table or another but not both in one query?

    I have a situation where I need to link two tables together but the data may be in another (archive) table or different records are in both but I want the latest record from either table:
    ACCOUNT
    AccountID     Name   
    123               John Doe
    124               Jane Donaldson           
    125               Harold Douglas    
    MARKETER_ACCOUNT
    Key     AccountID     Marketer    StartDate     EndDate
    1001     123               10526          8/3/2008     9/27/2009
    1017     123               10987          9/28/2009     12/31/4712    (high date ~ which means currently with this marketer)
    1023     124               10541          12/03/2010     12/31/4712
    ARCHIVE
    Key     AccountID     Marketer    StartDate     EndDate
    1015     124               10526          8/3/2008     12/02/2010
    1033     125               10987         01/01/2011     01/31/2012  
    So my query needs to return the following:
    123     John Doe                        10526     8/3/2008     9/27/2009
    124     Jane Donaldson             10541     12/03/2010     12/31/4712     (this is the later of the two records for this account between archive and marketer_account tables)
    125     Harold Douglas               10987          01/01/2011     01/31/2012     (he is only in archive, so get this record)
    I'm unsure how to proceed in one query.  Note that I am reading in possibly multiple accounts at a time and returning a collection back to .net
    open CURSOR_ACCT
              select AccountID
              from
                     ACCOUNT A,
                     MARKETER_ACCOUNT M,
                     ARCHIVE R
               where A.AccountID = nvl((select max(M.EndDate) from Marketer_account M2
                                                    where M2.AccountID = A.AccountID),
                                                      (select max(R.EndDate) from Archive R2
                                                    where R2.AccountID = A.AccountID)
                   and upper(A.Name) like parameter || '%'
    <can you do a NVL like this?   probably not...   I want to be able to get the MAX record for that account off the MarketerACcount table OR the max record for that account off the Archive table, but not both>
    (parameter could be "DO", so I return all names starting with DO...)

    if I understand your description I would assume that for John Dow we would expect the second row from marketer_account  ("high date ~ which means currently with this marketer"). Here is a solution with analytic functions:
    drop table account;
    drop table marketer_account;
    drop table marketer_account_archive;
    create table account (
        id number
      , name varchar2(20)
    insert into account values (123, 'John Doe');
    insert into account values (124, 'Jane Donaldson');
    insert into account values (125, 'Harold Douglas');
    create table marketer_account (
        key number
      , AccountId number
      , MktKey number
      , FromDt date
      , ToDate date
    insert into marketer_account values (1001, 123, 10526, to_date('03.08.2008', 'dd.mm.yyyy'), to_date('27.09.2009', 'dd.mm.yyyy'));
    insert into marketer_account values (1017, 123, 10987, to_date('28.09.2009', 'dd.mm.yyyy'), to_date('31.12.4712', 'dd.mm.yyyy'));
    insert into marketer_account values (1023, 124, 10541, to_date('03.12.2010', 'dd.mm.yyyy'), to_date('31.12.4712', 'dd.mm.yyyy'));
    create table marketer_account_archive (
        key number
      , AccountId number
      , MktKey number
      , FromDt date
      , ToDate date
    insert into marketer_account_archive values (1015, 124, 10526, to_date('03.08.2008', 'dd.mm.yyyy'), to_date('02.12.2010', 'dd.mm.yyyy'));
    insert into marketer_account_archive values (1033, 125, 10987, to_date('01.01.2011', 'dd.mm.yyyy'), to_date('31.01.2012', 'dd.mm.yyyy'));
    select key, AccountId, MktKey, FromDt, ToDate
         , max(FromDt) over(partition by AccountId) max_FromDt
      from marketer_account
    union all
    select key, AccountId, MktKey, FromDt, ToDate
         , max(FromDt) over(partition by AccountId) max_FromDt
      from marketer_account_archive;
    with
    basedata as (
    select key, AccountId, MktKey, FromDt, ToDate
      from marketer_account
    union all
    select key, AccountId, MktKey, FromDt, ToDate
      from marketer_account_archive
    basedata_with_max_intervals as (
    select key, AccountId, MktKey, FromDt, ToDate
         , row_number() over(partition by AccountId order by FromDt desc) FromDt_Rank
      from basedata
    filtered_basedata as (
    select key, AccountId, MktKey, FromDt, ToDate from basedata_with_max_intervals where FromDt_Rank = 1
    select a.id
         , a.name
         , b.MktKey
         , b.FromDt
         , b.ToDate
      from account a
      join filtered_basedata b
        on (a.id = b.AccountId)
    ID NAME                     MKTKEY FROMDT     TODATE
    123 John Doe                  10987 28.09.2009 31.12.4712
    124 Jane Donaldson            10541 03.12.2010 31.12.4712
    125 Harold Douglas            10987 01.01.2011 31.01.2012
    If your tables are big it could be necessary to do the filtering (according to your condition) in an early step (the first CTE).
    Regards
    Martin

  • Can not insert/update data from table which is created from view

    Hi all
    I'm using Oracle database 11g
    I've created table from view as the following command:
    Create table table_new as select * from View_Old
    I can insert/update data into table_new by command line.
    But I can not Insert/update data of table_new by SI Oject Browser tool or Oracle SQL Developer tool .(read only)
    Anybody tell me, what's happend? cause?
    Thankyou
    thiensu
    Edited by: user8248216 on May 5, 2011 8:54 PM
    Edited by: user8248216 on May 5, 2011 8:55 PM

    I can insert/update data into table_new by command line.
    But I can not Insert/update data of table_new by SI Oject Browser tool or Oracle SQL Developer tool .(read only)so what is wrong with the GUI tools & why posting to DATABASE forum when that works OK?

  • Not able to refresh the data in a table

    Hi In my application i fill data in a table on clikc of a button ..
    Following are the line of code i have user
    RichColumn richcol = (RichColumn)userTableData.getChildren().get(0);                                           (fetch the first column of my table)
    richcol.getChildren().add(reportedByLabel);                                                                   (reportedByLabel is a object of RichInputText)
    AdfFacesContext adfContext1 = AdfFacesContext.getCurrentInstance();
    adfContext1.addPartialTarget(richcol);
    adfContext1.addPartialTarget(userTableData);
    But on submit of that button table data is not refreshed after adding partial trigger on that table as well as that column also .... any idea??
    Edited by: Shubhangi m on Jan 27, 2011 3:50 AM

    Hi,
    The Code that you have shown adds an additional inputText component to the first column of a table.
    Is that your intention?
    If yes, please use the following code snippet to achieve your functionality:
    JSPX Code:
    <af:form id="f1">
    <af:commandButton text="Add Column" id="cb1"
    actionListener="#{EmployeesTableBean.onAddColumn}"/>
    <af:table value="#{bindings.Employees.collectionModel}" var="row"
    rows="#{bindings.Employees.rangeSize}"
    emptyText="#{bindings.Employees.viewable ? 'No data to display.' : 'Access Denied.'}"
    fetchSize="#{bindings.Employees.rangeSize}"
    rowBandingInterval="0"
    selectedRowKeys="#{bindings.Employees.collectionModel.selectedRow}"
    selectionListener="#{bindings.Employees.collectionModel.makeCurrent}"
    rowSelection="single" id="t1"
    binding="#{EmployeesTableBean.table}">
    <af:column sortProperty="EmployeeId" sortable="false"
    headerText="#{bindings.Employees.hints.EmployeeId.label}"
    id="c1">
    <af:outputText value="#{row.EmployeeId}" id="ot2">
    <af:convertNumber groupingUsed="false"
    pattern="#{bindings.Employees.hints.EmployeeId.format}"/>
    </af:outputText>
    </af:column>
    <af:column sortProperty="FirstName" sortable="false"
    headerText="#{bindings.Employees.hints.FirstName.label}"
    id="c2">
    <af:outputText value="#{row.FirstName}" id="ot1"/>
    </af:column>
    <af:column sortProperty="LastName" sortable="false"
    headerText="#{bindings.Employees.hints.LastName.label}"
    id="c3">
    <af:outputText value="#{row.LastName}" id="ot3"/>
    </af:column>
    </af:table>
    </af:form>
    Bean:
    public class EmployeesTableBean {
    private RichTable table;
    public EmployeesTableBean() {
    public void setTable(RichTable table) {
    this.table = table;
    public RichTable getTable() {
    return table;
    public void onAddColumn(ActionEvent actionEvent) {
    RichInputText newRichInputText = new RichInputText();
    newRichInputText.setId("new");
    newRichInputText.setValue("Name:");
    RichColumn richcol = (RichColumn)table.getChildren().get(0);
    richcol.getChildren().add(newRichInputText);
    AdfFacesContext adfContext1 = AdfFacesContext.getCurrentInstance();
    adfContext1.addPartialTarget(table);
    Thanks,
    Navaneeth

  • Unable to capture the adf table column sort icons using open script tool

    Hi All,
    I am new to OATS and I am trying to create script for testing ADF application using open script tool. I face issues in recording two events.
    1. I am unable to record the event of clicking adf table column sort icons that exist on the column header. I tried to use the capture tool, but that couldn't help me.
    2. The second issue is I am unable to capture the panel header text. The component can be identified but I was not able to identify the supporting attribute for the header text.

    Hi keerthi,
    1. I have pasted the code for the first issue
    web
                             .button(
                                       122,
                                       "/web:window[@index='0' or @title='Manage Network Targets - Oracle Communications Order and Service Management - Order and Service Management']/web:document[@index='0' or @name='1824fhkchs_6']/web:form[@id='pt1:_UISform1' or @name='pt1:_UISform1' or @index='0']/web:button[@id='pt1:MA:0:n1:1:pt1:qryId1::search' or @value='Search' or @index='3']")
                             .click();
                        adf
                        .table(
                                  "/web:window[@index='0' or @title='Manage Network Targets - Oracle Communications Order and Service Management - Order and Service Management']/web:document[@index='0' or @name='1c9nk1ryzv_6']/web:ADFTable[@absoluteLocator='pt1:MA:n1:pt1:pnlcltn:resId1']")
                        .columnSort("Ascending", "Name" );
         }

  • How to Activate the Table which Doesn't Exist - Can I create Table in SE11

    Hi Gurus,
    I am a BASIS Person ..so, I don't have much idea about ABAP Developement side . But, while applying patches to one of  the SAP Component, I get the following error message ( usr\sap\trans\log ) :
    Phase Import_Proper
    >>>
    SAPAIBIIP7.BID
    1 ED0301 *************************************************************************
    1 EDO578XFollowing objects not activated/deleted or activated/deleted w. warning:
    1EEDO519X"Table" "WRMA_S_RESULT_DATA" could not be activated
    1EEDO519 "Table Type" "WRMA_TT_UPDTAB_DSO" could not be activated
    1 ED0313 (E- Row type WRMA_S_RESULT_DATA is not active or does not exist )
    2WEDO517 "Domain" "WRMA_KAPPL" was activated (error in the dependencies)
    2WEDO517 "Data Element" "/RTF/DE_FISCPER" was activated (error in the dependencies)
    2WEDO517 "Data Element" "/RTF/DE_FISCVARNT" was activated (error in the dependencies)
    2WEDO517 "Data Element" "WRMA_DE_RCLASV" was activated (error in the dependencies)
    2WEDO517 "Data Element" "WRMA_DE_RMA_AMNT" was activated (error in the dependencies)
    2WEDO517 "Data Element" "WRMA_DE_RMA_INV" was activated (error in the dependencies)
    2WEDO517 "Data Element" "WRMA_DE_RMA_OBJ" was activated (error in the dependencies)
    2WEDO549 "Table" "V_WRMA_TRCLASVER" was activated (warning for the dependent tables)
    2WEDO549 "Table" "V_WRMA_TRCLASVPE" was activated (warning for the dependent tables)
    2WEDO549 "Table" "V_WRMA_TRCLPERCL" was activated (warning for the dependent tables)
    2WEDO549 "Table" "WRMA_S_STOCK_SELECTION_DATE" was activated (warning for the dependent tables)
    <<<
    Now, I checked Domain WRMA_KAPPL  in SE11 and it diplays that it is partially active :
    " The obeject is marked partially active for the following reasons :
       Errors occured when adjusting dependent Objects to a change. Some of the dependent objects could not
       be adjusted to this change. To adjust the dependent objects, you must perform the following actions
       the next time you activate this object.
                        - SET THE SYNP TIME STAMP IN THE RUNTIME OBJECT OFALL DEPENDENT OBJECTS
    My question is - how to activate this SYNP Time Stamp.. ?
                          - Once I activate this SYNP Time Stamp.. How do I activate this ..in SE11 or some other
                            TCode..
    Any / ALL Help is highy appreciated and would be rewarded with appropriate Reward Points.
    Thanks,
    Regards,
    - Ishan
    >> Ok, Now I forcefully activated the Domain and then all the 5  Data Elements in SE11 ...now I have error
         that table  WRMA_S_RESULT_DATA  and WRMA_TT_UPDTAB_DSO Could not be activated...
         ..And in SE11 > Input the first / Second Table  > Excute > Output : Table Does not Exist !!!
          - Now, How do I activate something which does not exist in the first place ?
        Any Input about this ?
    Thanks,
    Regards,
    - Ishan
    Edited by: ISHAN P on Jun 19, 2008 3:44 PM

    Read note 1152612 - Incorrect component type in structure WRMA_S_RESULT_DATA.
    The error occurs in Release BI_CONT 703 Support Package 9.
    If you require an advance correction for Support Package 9, make the following manual changes if the InfoObject 0TCTLSTCHG is inactive in your system.
    1. Use transaction RSA1 to call the Data Warehousing Workbench.
    2. Check whether the InfoObject 0TCTLSTCHG was already activated by the compilation process the first time you called transaction RSA1.  You can use transaction RSD1 to check this by entering 0TCTLSTCHG in the "InfoObject" field and choosing "Display".  Proceed as follows if the InfoObject is still not active:
    3. Go to BI Content activation.
    4. Select "Object Types".
    5. Open the "InfoObjects" tree and select "Objects".
    6. Search for the InfoObject 0TCTLSTCHG and select it.
    7. Transfer the object and activate it.
    These steps activate the required component type /BI0/OITCTLSTCHG. You can check again in transaction RSD1 to ensure that the InfoObject 0TCTLSTCHG is available as active.
    Jacek Fornalczyk

Maybe you are looking for