Data Source Assignment issue in BW

Hi
i have replicate 2 data source in Production BW system but those are not automatically assigned to the corresponding
info source.Since Production system i can't do any change. it seems to me that i need to transport these 2 info source again in Production BW system.
Let me know your views.what needs to be done here to assign data source to Info si
Regards
Atul

Hi,
Before you transport data source / info-source to production system, you should make sure that in target system ( in your case production system ) you have maintained a setting of 'Mapping of Source system names'. This you can find in RSA1 --> Tools --> Mapping of source system names.
Here you define source system for data source should change to XXX(production) source system instead of earlier XXX(development) source system.
Regards,
Akshay

Similar Messages

  • Sharepoint 2013 Excel External Data Source Refresh Issue

    I have been facing this issue for quite some time now.. i have created an Excel sheet in Excel-13 and have imported data from an external data source [SQL server 2012]. 
    Everything is working fine, with the excel sheet on the desktop. Data refreshes, every-time i open the excel file and also at regular intervals that i have configured in the data source properties.
    The problem begins when i save that excel sheet on my sharepoint server. the issues that i am facing are :
    1. Changes made into the original data source, are not reflected immediately inside the excel sheet inside the browser. after 5-10 minutes, it reflects the changes..
    2. The data doesn't refreshes automatically. After i update my data inside the sql server table, i have to manually trigger the refresh of the data connection when viewing the excel sheet inside the browser, even though i have marked "Refresh when opening
    the file", and refresh every 1 minute inside the excel sheet. Any solutions ??
    I have been troubled a lot by this issue, and seek for some quick solution.. Any help here ??

    I found the solution finally, my self ..
    Issue - 1 : It's going to take atleat 5-minutes to refresh the data connection, that is generally not a big time span.
    Issue - 2 : 
    --> Set Your connection to refresh everytime the file is opened. go to internet explorer -> file -> internet options -> general -> Browsing History -> Settings -> Check for newer versions of stored pages... Check 'Every time I visit the
    webpage'. 
    Now everytime i update your original data source, wait for 5-10 minutes and refresh my web page containing the excel sheet.. The Contents of the excel sheet are updated as desired..

  • SSRS 2008 R2 data source authentication issue

    Hello,
    End-users with Browser permissions for the necessary SSRS folders and reports are getting "The permissions denied to user [Windows user] are insufficient for performing this operation (rsAccessDenied)" even though the data source for the report is configured
    to use "credentials stored securely in the report server" and "Use as Windows credentials when connecting to the data source" is checked. Users who are members of the local admin machine hosting the Report Manager site don't get the error. But users with full
    control permissions to the SSRS directories, including the config files, do get the error. Obviously, the solution shouldn't entail giving end-users local admin or even full control permissions to the SQL Server SSRS folders.  Moreover, the error only
    occurs when using Windows credentials but not when using only a database credential, i.e. non Windows credentials.  This workaround doesn't work for me since I'm also using a data source connection to SSAS, which doesn't appear to use database logins
    - only Windows credentials that are added to roles.
    Please advise.
    Thanks,
    Ben Lezin

    Hi Ben,
    Generally, the error "The permissions denied to user [Windows user] are insufficient for performing this operation (rsAccessDenied)" should be caused
    by one of the following reasons:
    1. User Access Control (UAC). Windows Vista, Windows 7 and Windows Server 2008 limit the overuse of elevated permissions by removing administrator permissions
    when accessing applications. Because the operating system removes permissions, members of the local Administrators group run most applications as if they are using the Standard User account.
    If this issue is caused by UAC, please run Internet Explorer as administrator. From the Start menu, click All Programs, right-click Internet Explorer,
    and select Run as administrator. For more information about UAC, please refer to this document:
    http://msdn.microsoft.com/en-us/library/bb630430.aspx
    2. User has insufficient permissions to perform the operation. Please check the following items:
    a. End-users belong to "Browser" role. In Report Manager Home page, click "Folder Settings", make sure the end-users are in the security list and with
    Browser role permission.
    b. "Browser" role has permissions to perform the operation. You can use SSMS to connect to report server, expand Security/Roles, double click "Browser",
    and then check the tasks list.
    If there is anything unclear, please feel free to ask.
    Thanks,
    Albert Ye

  • Data Source assignment problem

    Hi All,
    I have a Datasource1 assigned to Infosource1. DS1-->IS1. working fine.
    No I want to assign Data Source1 to infosource2, DS1-->IS2. ??
    In IS2-> right click--> selected Sourcesystem. But I can't see the DS1 in the list. I thought it was already assigned to IS1?? Please correct me if I am wrong?
    Is it possible to the same data source to assign multiple Infosources? If yes, how?
    Please advise.
    Thanks In advance.
    Sri.

    Hi,
    You cannot assign the same datasource to more than one info source.
    Instead create more update rules to your new ODS/CUBE for the same Info source
    Good Luck

  • ODP Data Source Delta issue 2lis_11_VAHDR 2lis_11_VAKON

    Hi Expert,
    I am pulling the data from SAP using ODP data sources. 2LIS_11_VAHDR 2LIS_11_VAITM 2LIS_11_VAKON.
    For all the data sources I am doing the initloads. For 2LIS_11_VAITM data source after init load its enabling the delta.
    For remaining data sources even if you try to pull the delta it's loading full load.
    Steps I am following
    LBWG do the deletion of setup tables.
    OLI7BW fill the setup tables.
    Run the data services job which pull the initial load. It loads all the data.
    Change the ODP data source settings to Init load to No (Capture only delta)
    In SMQ1 all the delta records will be captured
    Run the Jobs which pull the delta records to ODQMON from there if you run Job in data services it has to pull the Delta records.
    This is happening for 2LIS_11_VAITM but not for other data sources 2LIS_11_VAHDR and 2LIS_11_VAKON.
    Event if you try to pull the delta records its loading all the records.
    Can some one throw some light on this
    Regards,
    Murali.

    In RSMO, go to the details tab and check in which stage the load has failed. If the issue is with extraction then go to Environment > Job Overview in the source system. Check if this job is complete.
    I hope you must have cleared the Delta Queue, so now go and check if any entry is being displayed for 2lis_11_vahdr

  • JDBC Data Sources: Potential Issue with JDeveloper 10.1.3.4

    I think I found a bug or issue with the latest JDeveloper 10.1.3.4 release when using JDBC Data Sources on the Embedded OC4J container.
    To state the issule bluntly, If I use a JDBC Data Source in an ADF Faces application, I get the following error on the screen when I run my application if I create a simple page using a Form layout for database data:
    [http://cs.uwindsor.ca/~ruston7/jdbcError.jpg]
    Or if I use a simple drag and drop ADF Faces Table:
    javax.faces.el.PropertyNotFoundException: Error testing property '<<FIRST_FIELD_ON_THE_PAGE>>' in bean of type null
        at com.sun.faces.el.PropertyResolverImpl.isReadOnly(PropertyResolverImpl.java:274)
        at oracle.adfinternal.view.faces.model.FacesPropertyResolver.isReadOnly(FacesPropertyResolver.java:124)
        at com.sun.faces.el.impl.ArraySuffix.isReadOnly(ArraySuffix.java:236)
        at com.sun.faces.el.impl.ComplexValue.isReadOnly(ComplexValue.java:209)
        at com.sun.faces.el.ValueBindingImpl.isReadOnly(ValueBindingImpl.java:266)
        at oracle.adfinternal.view.faces.renderkit.core.xhtml.EditableValueRenderer.getReadOnly(EditableValueRenderer.java:211)
        at oracle.adfinternal.view.faces.renderkit.core.xhtml.FormElementRenderer.renderAsElement(FormElementRenderer.java:155)
        at oracle.adfinternal.view.faces.renderkit.core.xhtml.InputLabelAndMessageRenderer.getLabelFor(InputLabelAndMessageRenderer.java:53)
        at oracle.adfinternal.view.faces.renderkit.core.xhtml.LabelAndMessageRenderer$Label.getForId(LabelAndMessageRenderer.java:500)
        at oracle.adfinternal.view.faces.renderkit.core.xhtml.OutputLabelRenderer.encodeAll(OutputLabelRenderer.java:69)
        at oracle.adfinternal.view.faces.renderkit.core.CoreRenderer.delegateRenderer(CoreRenderer.java:281)
        at oracle.adfinternal.view.faces.renderkit.core.xhtml.LabelAndMessageRenderer.encodeAll(LabelAndMessageRenderer.java:123)
        at oracle.adfinternal.view.faces.renderkit.core.xhtml.InputLabelAndMessageRenderer.encodeAll(InputLabelAndMessageRenderer.java:94)
        at oracle.adfinternal.view.faces.renderkit.core.CoreRenderer.encodeEnd(CoreRenderer.java:169)
        at oracle.adf.view.faces.component.UIXComponentBase.encodeEnd(UIXComponentBase.java:624)
        at oracle.adfinternal.view.faces.renderkit.core.CoreRenderer.encodeChild(CoreRenderer.java:246)When I change my Application Module connection to a JDBC URL, this all works perfectly. Also, everything works fine when I deploy to our Oracle AS 10.1.3 application servers.
    I also tried this on a different computer using a fresh install of JDeveloper just to make sure that the copy of JDeveloper that I downloaded didn't have a fluke in it.
    Thanks!

    M. Ruston,
    It must be something on your side. i just tried the same thing using the employees table from the HR sample schema (it has a date column). It works with JDBC URL and datasource both.
    Just out of curiosity - if you look at the properties for your Business Components project at the business components section, what does it show for the SQL Flavor and Type Map?
    John

  • CRM Data source 0CRM_OPPT_H  issue with full update

    Hi Gurus!
    When I tried to start Full Upload via 0CRM_OPPT_H data source at BW side  I faced with issue that
    No data available
    Diagnosis
    The data request was a full update.
    In this case, the corresponding table in the source system does not
    contain any data.
    System Response
    Info IDoc received with status 8.
    Procedure
    Check the data basis in the source system.
    I have checked the data source via RSA3 at CRM side and can extract 11 entries. The table CRMD_ORDER_INDEX has 39 entries.
    What have I done:
    1. Activate data source at CRM side in RSA5 and can see it in RSA6
    2. Activate BW Adapter Metadata in BWA5
    3. Tried to start full uploading at BW side but got error(No data available).
    4. Initialized delta-upload and can see active status for 0CRM_OPPT_H in tr. BWA7 at CRM side in active status (Delta active = TRUE, Initial upload = empty).
    5. There is no any authorization issues (user has SAP_ALL in both systems)
    6. RFC connection between systems is Ok. Extraction via 0CRM_OPPT_I was Ok.
    Have you any ideas how to start full uploading for BW via 0CRM_OPPT_H?

    Please, look above:
    "I have checked the data source via RSA3 at CRM side and can extract 11 entries. The table CRMD_ORDER_INDEX has 39 entries.

  • WWW Data Source/Proxy issue

    I've seen a similar post, but wanted to see if anyone has encountered either of these errors while trying to crawl external web sites into the KD:
    Job Operation #1 failed: This crawl could not be launched because the location from which it was supposed to start, http://www.cnn.com could not be found or was inaccessible. When the crawler attempted to visit this location it received the following message: Exception of type com.plumtree.openkernel.exceptions.OKAssertError was thrown.(282610)
    Job Operation #1 failed: This crawl could not be launched because the location from which it was supposed to start, http://www.cnn.com/ could not be found or was inaccessible. When the crawler attempted to visit this location it received the following message: -2147203840 - Error in function PTWebCrawlProvider.GetChildNodes (pSafeArrayDocuments == [email protected], pSafeArrayContainers == [email protected]): -2147203840 - CPTWebCrawlProvider::GetMIMEType, could not open : http://www.cnn.com/(282610)
    We have the proxy info in the www data source, and the serverconfig.xml seems ok, but we are getting 407 (proxy) http responses.
    Thanks,
    Daryl

    In addition to Mti post..
    You said you had enhanced the DS. have written Code in the CMOD to populate the Data to the fileds.
    If already written check the tick mark "Filed only know for Exit". the system will come to know that data is being populated using the FM.

  • *** Transporting Generated Export Data Source *** urgent issue

    Hi All,
    I have generated export data source from info object Company Code,
    And trying to transport the Export data source, i.e. 8Company_code to QA system.
    Its showing the log that, imported sucessfully, but, i can't see the object in QA.
    Iam currently working on BW 3.5 system.
    Please help.
    Thanks,
    Nisha

    Hi Luis,
    Thanks for this.
    Yes, i went into Transport connector >> choosen the object >> Data flow before >>
    Ticked the Data sources >> changed the Package from $TMP to ZXXX >> Generated Transport request >> Released it >> Transported it.
    But, still can't see....
    Cheers,
    Nisha

  • Generic Data Source - Delta Issue

    HI,
    I have created a generic delta source (transactions) with the following options: -
    -  using the extract structure
    - enabled the generic delta with billing creation date as delta specific field
    - calendar day selected
    - safety interval upper limit = 1 calendar days
    - additive delta
    carried out the folling steps in BW
    - Full update from above generic data source (all records)
    - Init upload without data transfers (1 record)
    - Delta enabled
    Now, even though i am editing / modifying any record on my own extract strucutres record on R/3 side, it is not getting reflected in RSA7 (delta queue).
    P.S:- I have modified one of the record from full update.
    Ideally the modified record should be shown in RSA7, but i am not getting the same.
    can you please help me
    Thanks
    Raj

    yes,
    I also thought the same point. But the records modified / edited yesterday are not appearing in today's delta. I am wondering why.
    Is it anything to deal with FULL / INIT without deta transfer / DELTA procedure..
    Regards
    Rajiv

  • Data Source is in Inactive state while replicate datasource in BI 7.0

    Hello,
    When I replicate the datasource ZW_MAT_PLANT(My own data source which is combination of MARA and MARC) in RSA1-Datasources it is coming by default in inactive mode. In this inactive mode i can not delete this data source assignment adn i can not go inside this data source. While i click change icon to go inside the datasource it is giving me the error like you have no authorization for this datasource.
    But from SU53 screen shot i have authorization for this.
    Please help me in this.
    Thanks and Best Regards,
    Anitha Sukhaavsi

    Hi,
    I am facing the same issue. Did you get the solution for this?
    Thanks,
    Pal

  • Setting a data source for a tableView

    Greetings
    I'm trying to learn cocoa touch by following the Apple's code example called SQLiteBooks.
    there is a data source assignment within the MasterViewController code:
    - (void)loadView {
    tableView = [[[UITableView alloc] initWithFrame:[[UIScreen mainScreen] applicationFrame] style:UITableViewStylePlain] autorelease];
    tableView.delegate = self;
    tableView.dataSource = self;
    self.view = tableView; }
    Can anybody explain how "tableView.dataSource = self" associates an NSArray "books", which actually has the data, with this tableView?
    Best regards,
    -Eugene

    I can give you a Cocoa-specific answer. Unfortunately, with the new iPhone SDK there are tons of new people posting questions about the iPhone. That is why I didn't respond yesterday. My answer may or may not make sense for the iPhone.
    When you setup a data source, the table view will call two methods on that data source to get data. It will call "numberOfRowsInTableView:" to get the number of data elements and "tableView:objectValueForTableColumn:row:" to get a specific value for a cell.
    Your data source should implement those methods and return data for the table view to dipslay.
    Full details are available in Apple's documentation

  • Authorization Issue on creating New Data Source

    Hi Experts,
       We met several issues after we upgrade one of our BW system from 3.5 to BI 7.0.
       We have a customized role for our BW developer which name is SYN:BC_USER_FOR_BWDEVELOPER. When we assign this role to the users, the users should have access to create new data source, create/change on ODS,  infocube, infobjects,etc. The developers also need access transformation roles and BEX Web Templates.
       After we upgrade our BW system from 3.5 to BI7.0. All the access has been denied. Then I tried to added many authorization objects in the developer role, then the developer role can access all the data source except transformation rules.
      The following is the authorization objects which I have added:
    S_RS_BCS  
    S_RS_BITM
    S_RS_COMP
    S_RS_COMP1
    S_RS_CTT
    S_RS_ERPT
    S_RS_HIER
    S_RS_ICBE
    S_RS_IOBJ
    S_RS_IOMAD
    S_RS_ISET
    S_RS_ISOUR
    S_RS_ISRUM
    S_RS_THJT
    S_RS_TOOLS
    S_RS_UOM.
    S_RS_DS
    S_RS_ISNEW
    S_RS_DTP
    S_RS_TR
    S_RS_PC
    S_RS_OHDST
    S_RS_DAS
    S_RS_BTMP.
      I am not sure whether there are any other authorization objects or sth should be added to the developer role, could you give me some suggestions?
    Thanks,

    When i execute create transformation rules with the account which already have bw_all role, the system shows you have no authorization.
    The following is the error after i execute /nsu53:
    Authorization check failed.
    Authorization Obj. S_RS_TR Data Warehousing Workbench -Transformation
      Object Class RS Business Information Warehouse
       Activity                                03
       Source                                 ZM_SUM
       Target                                   ZWHYTEST
       Subtype of the Source            ‘ ’
       Subtype of the Target             ‘ ’
       Type of Source                      ODSO
       Type of Target                       TRCS
    -fran

  • Z data source issue for creating packets

    Hi I have created a Z data source (function module) .
    My issue is I am not able to create data record packets all the data is coming in one packet only.
    The is code is as show below is some one can please assist me how can I change the code so that is can create multiple packets for the option given in Tcode RSA3.
    FUNCTION ZBW_MATERIAL_GROUP_HIE.
    ""Local Interface:
    *" IMPORTING
    *" VALUE(I_REQUNR) TYPE SRSC_S_IF_SIMPLE-REQUNR
    *" VALUE(I_DSOURCE) TYPE SRSC_S_IF_SIMPLE-DSOURCE OPTIONAL
    *" VALUE(I_MAXSIZE) TYPE SRSC_S_IF_SIMPLE-MAXSIZE OPTIONAL
    *" VALUE(I_INITFLAG) TYPE SRSC_S_IF_SIMPLE-INITFLAG OPTIONAL
    *" VALUE(I_READ_ONLY) TYPE SRSC_S_IF_SIMPLE-READONLY OPTIONAL
    *" VALUE(I_REMOTE_CALL) TYPE SBIWA_FLAG DEFAULT SBIWA_C_FLAG_OFF
    *" TABLES
    *" I_T_SELECT TYPE SRSC_S_IF_SIMPLE-T_SELECT OPTIONAL
    *" I_T_FIELDS TYPE SRSC_S_IF_SIMPLE-T_FIELDS OPTIONAL
    *" E_T_DATA STRUCTURE ZBW_MAT_GRP_HIER OPTIONAL
    *" EXCEPTIONS
    *" NO_MORE_DATA
    *" ERROR_PASSED_TO_MESS_HANDLER
    TABLES : /BI0/HMATL_GROUP.
    DATA : BEGIN OF t_hmat OCCURS 0,
    hieid LIKE /BI0/HMATL_GROUP-hieid,
    objvers LIKE /BI0/HMATL_GROUP-objvers,
    iobjnm LIKE /BI0/HMATL_GROUP-iobjnm,
    nodeid LIKE /BI0/HMATL_GROUP-nodeid,
    nodename LIKE /BI0/HMATL_GROUP-nodename,
    tlevel LIKE /BI0/HMATL_GROUP-tlevel,
    parentid LIKE /BI0/HMATL_GROUP-parentid,
    END OF t_hmat.
    DATA : BEGIN OF t_flathier,
    hieid LIKE /BI0/HMATL_GROUP-hieid,
    lv2_id LIKE /BI0/HMATL_GROUP-nodeid,
    lv2_name LIKE /BI0/HMATL_GROUP-nodename,
    lv3_id LIKE /BI0/HMATL_GROUP-nodeid,
    lv3_name LIKE /BI0/HMATL_GROUP-nodename,
    lv4_id LIKE /BI0/HMATL_GROUP-nodeid,
    lv4_name LIKE /BI0/HMATL_GROUP-nodename,
    lv5_id LIKE /BI0/HMATL_GROUP-nodeid,
    lv5_name LIKE /BI0/HMATL_GROUP-nodename,
    lv6_id LIKE /BI0/HMATL_GROUP-nodeid,
    lv6_name LIKE /BI0/HMATL_GROUP-nodename,
    lv7_id LIKE /BI0/HMATL_GROUP-nodeid,
    lv7_name LIKE /BI0/HMATL_GROUP-nodename,
    lv8_id LIKE /BI0/HMATL_GROUP-nodeid,
    lv8_name LIKE /BI0/HMATL_GROUP-nodename,
    lv9_id LIKE /BI0/HMATL_GROUP-nodeid,
    lv9_name LIKE /BI0/HMATL_GROUP-nodename,
    lv10_id LIKE /BI0/HMATL_GROUP-nodeid,
    lv10_name LIKE /BI0/HMATL_GROUP-nodename,
    lv11_id LIKE /BI0/HMATL_GROUP-nodeid,
    lv11_name LIKE /BI0/HMATL_GROUP-nodename,
    material LIKE /BI0/HMATL_GROUP-nodename,
    END OF t_flathier.
    FIELD-SYMBOLS: <f> LIKE LINE OF t_hmat,
    <Level> TYPE ANY.
    data : count(2) type c,
    lv_level(20) type c.
    DATA : lv_count TYPE n.
    DATA : lv_id LIKE /BI0/HMATL_GROUP-nodeid,
    lv_hieid LIKE /BI0/HMATL_GROUP-hieid.
    Auxiliary Selection criteria structure
    DATA: l_s_select TYPE srsc_s_select.
    Maximum number of lines for DB table
    STATICS: s_s_if TYPE srsc_s_if_simple,
    counter
    s_counter_datapakid LIKE sy-tabix,
    cursor
    s_cursor TYPE cursor.
    Select ranges
    RANGES: l_r_nodename FOR /BI0/HMATL_GROUP-nodename,
    l_r_hieid FOR /BI0/HMATL_GROUP-hieid.
    Initialization mode (first call by SAPI) or data transfer mode
    (following calls) ?
    IF i_initflag = sbiwa_c_flag_on.
    Initialization: check input parameters
    buffer input parameters
    prepare data selection
    Check DataSource validity
    CASE i_dsource.
    WHEN 'ZMATERIAL_GROUP_HIE'.
    WHEN OTHERS.
    IF 1 = 2. MESSAGE e009(r3). ENDIF.
    this is a typical log call. Please write every error message like this
    log_write 'E' "message type
    'R3' "message class
    '009' "message number
    i_dsource "message variable 1
    ' '. "message variable 2
    RAISE error_passed_to_mess_handler.
    ENDCASE.
    APPEND LINES OF i_t_select TO s_s_if-t_select.
    Fill parameter buffer for data extraction calls
    s_s_if-requnr = i_requnr.
    s_s_if-dsource = i_dsource.
    s_s_if-maxsize = i_maxsize.
    Fill field list table for an optimized select statement
    (in case that there is no 1:1 relation between InfoSource fields
    and database table fields this may be far from beeing trivial)
    APPEND LINES OF i_t_fields TO s_s_if-t_fields.
    ELSE. "Initialization mode or data extraction ?
    Data transfer: First Call OPEN CURSOR + FETCH
    Following Calls FETCH only
    First data package -> OPEN CURSOR
    IF s_counter_datapakid = 0.
    Fill range tables BW will only pass down simple selection criteria
    of the type SIGN = 'I' and OPTION = 'EQ' or OPTION = 'BT'.
    LOOP AT s_s_if-t_select INTO l_s_select WHERE fieldnm = '0MATERIAL'.
    MOVE-CORRESPONDING l_s_select TO l_r_nodename.
    APPEND l_r_nodename.
    ENDLOOP.
    LOOP AT s_s_if-t_select INTO l_s_select WHERE fieldnm = 'HIEID'.
    MOVE-CORRESPONDING l_s_select TO l_r_hieid.
    APPEND l_r_hieid.
    ENDLOOP.
    Get the data from Hierarchy table
    SELECT * FROM /BI0/HMATL_GROUP INTO CORRESPONDING FIELDS OF
    TABLE t_hmat
    WHERE hieid IN l_r_hieid
    AND objvers = 'A' .
    ENDIF.
    loop through all the 0MATERIAL entries to get all the hirarchy levels.
    Start of change.
    LOOP AT t_hmat ASSIGNING <f>
    WHERE iobjnm = '0MATL_GROUP'
    AND nodename IN l_r_nodename.
    LOOP AT t_hmat ASSIGNING <f>
    WHERE nodename IN l_r_nodename.
    End of change
    lv_count = <f>-tlevel.
    "refresh t_flathier.
    CLEAR: t_flathier. ", lv_level, count.
    MOVE :
    <f>-hieid TO lv_hieid ,
    <f>-nodename TO t_flathier-material,
    <f>-parentid TO lv_id.
    if <f>-iobjnm <> '0MATL_GROUP' .
    move <f>-nodename+3 to t_flathier-material .
    else.
    move <f>-nodename to t_flathier-material .
    endif.
    Added for Last level.
    if lv_count = '1' .
    *t_flathier-lv1_name = t_flathier-material .
    elseif lv_count = '2' .
    t_flathier-lv2_name = t_flathier-material .
    elseif lv_count = '3' .
    t_flathier-lv3_name = t_flathier-material .
    elseif lv_count = '4' .
    t_flathier-lv4_name = t_flathier-material .
    elseif lv_count = '5' .
    t_flathier-lv5_name = t_flathier-material .
    elseif lv_count = '6' .
    t_flathier-lv6_name = t_flathier-material .
    elseif lv_count = '7' .
    t_flathier-lv7_name = t_flathier-material .
    elseif lv_count = '8' .
    t_flathier-lv8_name = t_flathier-material .
    elseif lv_count = '9' .
    t_flathier-lv9_name = t_flathier-material .
    elseif lv_count = '10' .
    t_flathier-lv10_name = t_flathier-material .
    endif.
    DO lv_count TIMES .
    lv_count = lv_count - 1.
    IF lv_count = 1.
    EXIT.
    ENDIF.
    READ TABLE t_hmat WITH KEY
    hieid = lv_hieid
    nodeid = lv_id.
    IF sy-subrc = 0.
    CLEAR lv_id.
    CASE lv_count.
    WHEN '11' .
    MOVE : t_hmat-nodename+3 TO t_flathier-lv11_name,
    t_hmat-parentid TO lv_id.
    WHEN '10' .
    MOVE : t_hmat-nodename+3 TO t_flathier-lv10_name,
    t_hmat-parentid TO lv_id.
    WHEN '9' .
    MOVE : t_hmat-nodename+3 TO t_flathier-lv9_name,
    t_hmat-parentid TO lv_id.
    WHEN '8' .
    MOVE : t_hmat-nodename+3 TO t_flathier-lv8_name,
    t_hmat-parentid TO lv_id.
    WHEN '7' .
    MOVE : t_hmat-nodename+3 TO t_flathier-lv7_name,
    t_hmat-parentid TO lv_id.
    WHEN '6' .
    MOVE : t_hmat-nodename+3 TO t_flathier-lv6_name,
    t_hmat-parentid TO lv_id.
    WHEN '5' .
    MOVE : t_hmat-nodename+3 TO t_flathier-lv5_name,
    t_hmat-parentid TO lv_id.
    WHEN '4' .
    MOVE : t_hmat-nodename+3 TO t_flathier-lv4_name,
    t_hmat-parentid TO lv_id.
    WHEN '3' .
    MOVE : t_hmat-nodename+3 TO t_flathier-lv3_name,
    t_hmat-parentid TO lv_id.
    WHEN '2' .
    MOVE : t_hmat-nodename+3 TO t_flathier-lv2_name.
    ENDCASE.
    ENDIF.
    ENDDO.
    Populate data for level 1 (Class Type)
    READ TABLE t_hmat WITH KEY
    hieid = lv_hieid
    tlevel = 1.
    IF sy-subrc = 0.
    CALL FUNCTION 'CONVERSION_EXIT_ALPHA_OUTPUT'
    EXPORTING
    input = t_hmat-nodename
    IMPORTING
    output = e_t_data-0class_type.
    CALL FUNCTION 'CONVERSION_EXIT_ALPHA_INPUT'
    EXPORTING
    input = e_t_data-0class_type
    IMPORTING
    output = e_t_data-0class_type.
    ENDIF.
    populate data to extraction structure ( removing prefixe 'class type')
    MOVE : lv_hieid TO e_t_data-hieid,
    t_flathier-lv2_name TO e_t_data-xhier_lv1,
    t_flathier-lv3_name TO e_t_data-xhier_lv2,
    t_flathier-lv4_name TO e_t_data-xhier_lv3,
    t_flathier-lv5_name TO e_t_data-xhier_lv4,
    t_flathier-lv6_name TO e_t_data-xhier_lv5,
    t_flathier-lv7_name TO e_t_data-xhier_lv6,
    t_flathier-lv8_name TO e_t_data-xhier_lv7,
    t_flathier-lv9_name TO e_t_data-xhier_lv8,
    t_flathier-lv10_name TO e_t_data-xhier_lv9,
    t_flathier-lv11_name TO e_t_data-xhie_lv10,
    t_flathier-material TO e_t_data-0MATL_GROUP.
    APPEND e_t_data.
    CLEAR e_t_data.
    ENDLOOP.
    s_counter_datapakid = s_counter_datapakid + 1.
    IF s_counter_datapakid > 1 .
    RAISE no_more_data.
    ENDIF.
    ENDIF. "Initialization mode or data extraction ?
    ENDFUNCTION.
    As now when I run it in Tcode RSA3 it give only one data packet of some 5k to 6k records.
    Thanks in advance for your help.
    Pawan.

    Hi PS,
    Instead of
    SELECT * FROM /BI0/HMATL_GROUP INTO CORRESPONDING FIELDS OF
    TABLE t_hmat
    WHERE hieid IN l_r_hieid
    AND objvers = 'A' .
    code should look like this .
          OPEN CURSOR WITH HOLD S_CURSOR FOR
          SELECT (S_S_IF-T_FIELDS) FROM /BI0/HMATL_GROUP
        FETCH NEXT CURSOR S_CURSOR
                   APPENDING CORRESPONDING FIELDS
                   OF TABLE E_T_DATA
                   PACKAGE SIZE S_S_IF-MAXSIZE.
    For more information refer to sample code of fm "RSAX_BIW_GET_DATA_SIMPLE"
    Hope that helps.
    Regards
    Mr Kapadia
    ***Assigning points is the way to say thanks in SDN.***

  • Issue with table ROOSPRMSF entries for data source 0FI_AP_4

    Hi Experts,
    I am facing with an issue where we found incosistencies with table ROOSPRMSF in R/3 system.
    In BW , we have done initializations based on fiscal period selections (none of the selections overlap) for data source 0FI_AP_4.
    We have done in total 7 initializations. So in BW system in table RSSDLINITSEL we have 7 initialization requests.
    But in R/3 system we have 49 records for data source 0FI_AP_4 in ROOSPRMSF table out of which 42 are invalid records.
    I suspect that these 42 invalid records are created due to the execution of program RSSM_OLTP_INIT_DELTA_UPDATE when the tables ROOSPRMSF are actually holding the 7 initialization request entries.   Due to this each and every initialization request is linked to rest of the other intialization requests and ended with 49 records in table ROOSPRMSF table.
    Now our data loads are running fine but daily a short dump is raised . In the daily loads, BW init records in RSSDLINITSEL are compared with ROOSPRMSF entries and all the other 42 records which are invalid are written into system log and a short dump is raised.
    In order to fix these inconsistencies i checked for OSS note 852443. (Point 3 in OSS note)
    But it is specified to delete the delta queue for data source 0FI_AP_4 in RSA7 and instructed to execute the program RSSM_OLTP_INIT_DELTA_UPDATE so that the ROOSPRMSF table will be reconstructed with valid records available in RSSDLINITSEL. 
    From OSS note 852443 point 3
    "3. If the RSSDLINIT table in the BW system already contains entries, check the requests listed there in the RNR column in the monitor (transaction RSRQ). Compare these entries with the entries in the ROOSPRMSF and ROOSPRMSC tables with the INITRNR field. If, in the ROOSPRMSF and ROOSPRMSC tables for your DataSource source system combination, there are more entries with different INITRNR numbers, use transaction RSA7 in an OLTP source system to delete all entries and then use the RSSM_OLTP_INIT_DELTA_UPDATE report mentioned in the next section. For a DataMart source system, delete the entries that you cannot find in the RSSDLINIT table using the procedure described above."
    My question is if we delete the delta queue in RSA7 then all the tables in R/3 (ROOSPRMSF, ROOSPRMSC, Time stamp table) and BW (RSSDLINITSEL, initialization requests will be deleted) will be cleared. Then how will the program RSSM_OLTP_INIT_DELTA_UPDATE  copy entries into ROOSPRMSF table in R/3 ?
    Could any one please clarify this ?
    Thanks
    Regards,
    Jeswanth

    Hi Amarnath,
    Did you unhide the new field in RSA6 and regenerated the DataSource?
    Often SAP will populate newly added fields (belonging to the same (set) of table(s) used for extraction) automatically (e.g. SAP uses 'move-corresponding' in it's extractor-code, or, in this case, reading all fields from the DD, FM BWFIU_TRANSFORM_FIELDLIST).
    If the DataSource looks fine to you and the field is still not populated in RSA3 you can't go without a user-exit.
    Grtx,
    Marco

Maybe you are looking for