0Employee_Text master data returns 0 records with Full load

Hi All
0Employee master data texts returns 0 records with full update. I checked PA0000 table and it has 22K records. I am not sure what is missing.
Thanks for your help in advance.
Latha

Hi again Latha,
I've been checking the extractor 0EMPLOYEE_TEXT (RSA3 in debbug mode) and it gets data from table view V_BW_PA0001_ENM check if you allready have entries in this database view (it's a view from PA0001 with select condition PA0001-SPRPS eq ''.
If you have entries here you should get them with the extractor.
The notes solves extraction problems with several objects (in our case was 0ORGUNIT).
I hope this helps,
Regards.
Carlos.

Similar Messages

  • Check what master data was carried with last load

    Hello all,
    I have an issue. I'm looking for a solution, so maybe you can help me :).
    I have the InfoObject A linked to InfoObject B , through a transformation.
    I also have a DTP.
    Running the DTP , I load master data from A to B.
    A and B already contained data, before my load.
    How can I check what master data was carried with my last load?
    (I don't have the chance to check by request, as if in the case of a cube , for example)
    Thank you.

    I agree like cube we cannot see the request wise data wise data .
    The alternate to this is you are updating the data through the DTP correct .
    Goto DTP maintenanve screen and click on the PSA and check the date wise data in that .
    I hope you must be awared of  which date data you need to see . So you can check from the DTP PSA screen .
    Assign marks is the best way to appreciate help

  • Master Data: 2 records with different FROM-TO dates

    Hi,
    I am extracting the master data from R/3. In R/3 the master data record is created with valid FROM: 01/01/1900 and valid TO: 12/31/9999. But when I extract the data into BW it is coming as 2 records one with 01/01/1000-12/31/1899 range and the other with 01/01/1900-12/31/9999.
    What settings/steps I need to change to get only one record for now and of course in future any change to the records will comes as different valid date range.
    When I schedule the info package, the update tab consist of Fixed time interval start date: 10/20/2005; end date: 12/31/9999.
    Time being to avoid this in the transfer rule of FROM date info object selected constant option with date 01/01/1900 as in R/3????
    Appreciate your help on this.
    Thanks.
    Sudha.

    Hi Nagesh,
    Thanks for reply.
    I have assigned one-to-one from R/3 to BW.
    In R/3 I have FROM and TO date and assigned these two fields to Transfer structure FROM and TO info objects.
    I have maintained Transfer rule only after getting two records into BW for every one entry in R/3. In the rule I just assigned constant date to avoid 2 records for every entry from r/3. So nothing to do with the routine.
    I want to see one record in BW for every entry coming from R/3. At present in the BW master data sitting as 2 records.
    Example:
    one record in R/3 with:
    FROM date: 01/01/1900; TO date: 12/31/9999.
    Inserted 2 records into BW with:
    FROM date: 01/01/1000; TO date: 12/31/1899.
    FROM date: 01/01/1900; TO date: 12/31/9999.
    Thanks.
    Sudha.

  • DAC - Can I run more then one Exection Plans with full load option enabled?

    Hello All,
    I have configured for subject Area sales – Orders, Backlog and Invoices (OBIEE)
    Now I want to run DAC to load the underline tables for subject Areas Financials – Revenue Analysis and Financials – GL Budget and Expenses, I have identified the relevant fact tables and built a Execution Plan in DAC.
    DAC - In which mode should I run the execution plan?
    With full load option enabled or disabled because already full load run is done for sales – Orders, Backlog and Invoices and this run should not disturb the existing data
    If I run with full load option enabled do it overrides the existing data for other subject area
    Please help me I am bit confused!
    Regards,
    SMA
    Edited by: SMA on Aug 18, 2009 3:36 PM

    No need of combining all the subject areas into one execution plan. Simply run the second ETL without checking the full load option. It will do the full load for the tables related to the newly added subject areas and incremental for the common tables between old and new subject areas.

  • How can I load data into table with SQL*LOADER

    how can I load data into table with SQL*LOADER
    when column data length more than 255 bytes?
    when column exceed 255 ,data can not be insert into table by SQL*LOADER
    CREATE TABLE A (
    A VARCHAR2 ( 10 ) ,
    B VARCHAR2 ( 10 ) ,
    C VARCHAR2 ( 10 ) ,
    E VARCHAR2 ( 2000 ) );
    control file:
    load data
    append into table A
    fields terminated by X'09'
    (A , B , C , E )
    SQL*LOADER command:
    sqlldr test/test control=A_ctl.txt data=A.xls log=b.log
    datafile:
    column E is more than 255bytes
    1     1     1     1234567------(more than 255bytes)
    1     1     1     1234567------(more than 255bytes)
    1     1     1     1234567------(more than 255bytes)
    1     1     1     1234567------(more than 255bytes)
    1     1     1     1234567------(more than 255bytes)
    1     1     1     1234567------(more than 255bytes)
    1     1     1     1234567------(more than 255bytes)
    1     1     1     1234567------(more than 255bytes)

    Check this out.
    http://download-west.oracle.com/docs/cd/B10501_01/server.920/a96652/ch06.htm#1006961

  • HR master data change export with Interface-Toolbox PU12

    Hello out there !
    I implemented a HR master data change export with PU12 at one of my customers.
    It works fine - with one little problem:
    It is only possible to export one period (actual or selected).
    Is there any possibility to export more than one period (also periods in the future) ?
    The request is to export also master data with validity start date in a future period.
    Thanks for any help.
    Greetings, Holger Mächtig

    Dear Holger,
    I am very curious about your problem last year. Is it possible to do a future export with PU12?
    I am now working with PU12 for the first time and I have set up a export file in PU12.
    But when I am doing an update without filling the name for the file layout I have the following error:
    "An error occurred when opening the "export file"
    When I am filling the file layout I have the following error:
    "An error occurred when writing to an export file" (Error Number E107).
    "File processing: end of file".
    When I am not doing an update the export file runs and seems right.
    Do you know what I am doing wrong?
    Thanks in advance!
    Kind regards,
    Yvette

  • Fetch data using structure with full working example

    does any body tell me how to fetch data using structure with full working example
    the structure name is RSTXT and the field is TXLINE
    the data in the form of text is entered from the functional side using t-code ME52N
    from there i have to fetch the data
    in smartform or in report

    using this code to get text from ME52N  this is a structure still not getting output  
    DATA:BEGIN OF TA_ROW occurs 0,
          TXZ01(1000) TYPE C,
          END OF TA_ROW.
      DATA:BEGIN OF IT_ROW OCCURS 0,
          TXZ01(1000) TYPE C,
          END OF IT_ROW.
       DATA: thread LIKE thead.
       DATA: headerid TYPE char24.
       DATA: it_text LIKE tline OCCURS 0 WITH HEADER LINE.
       data:wa_banfn like eban-banfn.
       thread-tdid = 'B01'.
       thread-tdname = headerid.
       thread-tdobject = 'EBAN'.
         select txz01 from eban into corresponding fields of table ta_row
          where banfn = wa_banfn.
       headerid = '  '.
       CALL FUNCTION 'READ_TEXT'
         EXPORTING
      CLIENT                        = SY-MANDT
           id                            = thread-tdid
           language                      = sy-langu
           name                          = thread-tdname
           object                        = thread-tdobject
      ARCHIVE_HANDLE
                                    = 0
      LOCAL_CAT                     = ' '
    IMPORTING
      HEADER                        =
         TABLES
    lines                         = it_text
        EXCEPTIONS
          id                            = 1
          language                      = 2
          name                          = 3
          not_found                     = 4
          object                        = 5
          reference_check               = 6
          wrong_access_to_archive       = 7
          OTHERS                        = 8.
         LOOP AT it_text.
         CLEAR ta_row.
       BREAK-POINT.
      FROM THIS POINT TEXT IS COMING
         ta_row-txz01 = it_text-tdline.
         APPEND ta_row TO it_row.
         ENDLOOP.

  • Problems with Full loads/Decreased query performance in Prod

    We have a table which serves as a base for a complex view. The table has roughly around 10 Million records and its a daily full load(I know, that delta loads are much better for handling large sets of data, but this information is very dynamic and with the business time constraints and project deliverables, the best decision was to do a Full load).
    This is the process we follow:
    > Drop Indexes (All columns individual indexes which are used inside the complex view as joins)
    > Truncate table
    > Load data
    > Recreate indexes.
    All the above steps are performed from SAP Dataservices Thru scripts and sql() function to execute the command, no manual intervention what so ever.
    After the job is successfully completed, the view doesn't refresh at all(It sits there forever). The same job when run across same volume of production data in Test environment performs much faster.
    Then, how do I refresh the view is manually log into SQL Developer, drop all the indexes on the parent table, and re-create all the indexes in the same order as Dataservices script. It performs very well after till the next load (the next morning).
    Any suggestions would be very helpful.
    My main question is why does it run faster, when I drop and recreate the indexes? and doesnt complete when the indexes are created by the SQL() from data services tool.
    Tried:
    Explain Plan(in dev, Test, Prod): Query cost vaired accross environments but returned results with same return times (In Production after manual Index creation)
    Tuning advisor (Only in Test) DBA evaluated it to be good.
    Thanks
    Nash
    DB Version Oracle 11.0.7
    Dataservices 3.2

    BluShadow and Harman
    Thank You!
    Im using a regular view, not a materialized view. and yes the query plan is completely different from Test and Production. In test the query was completely running on Hash based joins whereas in Production its using Nested Joins in the execution plan.
    Will try to gather statistics after the load and as per BluShadow, will look on the way of writing a function that makes a call to Database.
    Thank you all for taking sometime. I will try to test this out starting today and will extend tests over a couple of days.
    Regards
    Nash

  • Update master data in BW with Visual Composer

    Hi guys.
    I have an attribute of an info object that I need to give the users access to update.
    I would like to do it with visual composer.
    I would like the user to see all rows in a table kind of view and the user will update only the rows that need to be changed and than he/she will click on a button that will update the master data table in BW.
    I believe that I'll need to use a BAPI or a RFC for that? Does anyone know any good predefined ones for that?
    Thanks
    Shlomi

    That's not really true. In VC you have access to all remote enabled functions. We have created specific functions to update attributes.
    We are using it to allow our users to update their own master data instead of bugging the BI team.
    Here is an example of our code.
    Before I paste this in, you need to know a couple things.
    1. You can only update one record at a time (with my code)
    2. You have to pass back in the entire record, even if you are only changing one attribute of many.
    3. It automatically does the attribute change run, so you do not need to to it.
    This is a simple infoobject with a few attributes.
    First we created a Get function (which collects the data and shows it to the user.
    Then we gave them a edit button on the table that goes to a pop up.
    The user selects the record to change, then hits the edit button.
    When the popup  comes up the data is shown in a form, with a save and cancel button, fields that you do not wish users to see or change, hide them, but they all most be present on the form.
    The  when the user hits save, it send it to the below function. The name of the infoobject is ZCAT_SUB.
    FUNCTION zcat_update_subparty2.
    ""Local Interface:
    *"  IMPORTING
    *"     VALUE(I_T_ZSUBPARTY) LIKE  /BIC/PZCAT_SUB
    *"  STRUCTURE  /BIC/PZCAT_SUB
    *"  EXPORTING
    *"     VALUE(E_RESULT) TYPE  STRING
      TYPE-POOLS: rs, rsarc, rsarr, sbiwa, rssm.
      DATA:
            l_t_zsubparty LIKE /bic/pzcat_sub OCCURS 10,
            l_s_minfo      TYPE  rssm_s_minfo,
            l_t_idocstate  TYPE  rsarr_t_idocstate,
            l_subrc        LIKE  sy-subrc.
      APPEND i_t_zsubparty TO l_t_zsubparty.
      CALL METHOD cl_rsdmd_update_master_data=>update_attributes_static
        EXPORTING
          i_iobjnm      = 'ZCAT_SUB'
          i_t_data      = l_t_zsubparty
          i_s_minfo     = l_s_minfo
          i_monitoring  = rs_c_false
        IMPORTING
          e_subrc       = l_subrc
        CHANGING
          c_t_idocstate = l_t_idocstate.
      IF sy-subrc <> 0.
        e_result = 'Error happened'.
      ELSE.
        e_result = 'Process Completed'.
        DATA:
          l_t_msg                  TYPE rs_t_msg,
          l_r_activate_master_data TYPE REF TO cl_rsdmd_activate_master_data.
        CREATE OBJECT l_r_activate_master_data
          EXPORTING
            i_iobjnm               = 'ZCAT_SUB'
            i_p_q_have_to_exist    = rs_c_true
          EXCEPTIONS
            iobj_not_found         = 1
            iobj_not_active        = 2
            generate_program_error = 3
            p_q_do_not_exist       = 4
            p_q_txt_do_not_exist   = 5
            OTHERS                 = 6.
        IF sy-subrc = 0.
          l_r_activate_master_data->activate( IMPORTING e_subrc = l_subrc
                                            CHANGING  c_t_msg = l_t_msg ).
        ENDIF.
      ENDIF.
    ENDFUNCTION.

  • Master data  deletion -  text with compounding

    HI All,
    I have " 0PYMT_METH" Info object which is having "0PYMT_METH_TEXT" Data source.  " 0PYMT_METH" Info object is having text and  0country with compounding attribute.
    - First , I did not load this master data , I loaded transactional data. so some values for key are loaded without text and country in master data.
    - Later i came to know we need text for paymnt method, then i loaded again master data. but i am not able to delete previous entries which have blank text.
    -i deleted data with SIDS and deleted all the transactional data , still i am getting that entries with blanks.Because of this in report output i am getting # value for the text.
    How to solve this? Is it because of compounding attribute ? If so can u please suggest regarding this ASAP?
    Thanks in advance,
    Jo

    Hi,
    Please check the data load for the text as you must be loading
    MAster Data CHAR , Country , LANGUAGE and TEXT as fields in your data source.
    As the Text can be over written so if the porper key fields are matched the text should be modifyed everytime.
    SO please check what combination you r  loading .
    ANyway if that is one time , and very less data you can manually maintain th text for master data .
      In order to delete teh data you use SE14 use the text table there to delete the data. PS : This is not recommended solution .

  • Master data check not working when loading into and ODS object

    Hi guys,
    we are loading transactional data into an ODS from a flat file.
    In the infopackage, we have checked the option :                                                
    "Do not update data if no master data exists for a characteristic".
    This option seems to be unreliable :
    For some characteristics (it seems those loaded using transfer rules (routines)), incorrect data is still loaded into the ODS even if no entry exist in the associated infoobject master data table.
    Any help appreciated.
    Many thanks !
    Pierre-Yves

    Hi Dear,
       please delete all the incorrect master data from the master data characteristic first and then try to uploading the data again .It should work, because there migth be some master data created when data were loaded with out the check before.
    And please check whether all your characteristic info object are of master data , I mean that in the infoobject defination you have checked the master data check box on the info object. If you have only text then it might fail the master data check.
    hope this help.
    cheers,
    balaji
    Message was edited by: Balaji NS

  • Characteristics - Master data Tab- infosource with direct update

    Dear Experts,
    Please explain me in Laymans Language as i am new to BI (but am an ABAP consultant).
    In Info Object Characteristics,  Master data Tab there is infosource with direct update with an application Component.
    1. Now my doubt is,  when i specify the Application Component, this infosource would appear under the application component in Modeling->Infosources.
    Then on this infosource i would right click and create the info pakage and then data transfer process.
    is it correct?
    2. Suppose for this characteristic if i dont specify the infosource with direct update  then it wont appear in the infosource, so how can i load data in this case?
    Regards
    BI Learner

    Hi,
    Using an InfoSource with direct updating, master data (characteristics with attributes, texts, or hierarchies) of an InfoObject can be updated directly (without update rules, only using transfer rules) into the master data table. To do this you must assign it an application component. The system displays the characteristic in the InfoSource tree in the Data Warehousing Workbench. You can assign DataSources and source systems to the characteristic from there. You can then also load master data, texts, and hierarchies for the characteristic.
    You cannot use an InfoObject as an InfoSource with direct updating if:
    i. The characteristic you want to modify is characteristic 0SOURSYSTEM (source system ID).
    ii.  The characteristic has neither master data nor texts nor hierarchies. It is therefore impossible to load data for the characteristic.
    iii.  The characteristic that you want to modify turns out not to be a characteristic, but a unit or a key figure.
    To generate an export DataSource for a characteristic, the characteristic must also be an InfoSource with direct updating.
    1. *Now my doubt is, when i specify the Application Component, this infosource would appear under the application component in Modeling->Infosources.
    Then on this infosource i would right click and create the info pakage and then data transfer process.*
    The characteristic will be available in info source tree, you can assign a datasource->assignsource system->create package->load data.
    2. Suppose for this characteristic if i dont specify the infosource with direct update then it wont appear in the infosource, so how can i load data in this case?
    For a characteristic If you dont specify 'InfoSources with Direct Updating' then it wont appear in infosource tree.

  • Cannot view Master Data Services Models with Excel Add-In, using MSSQServer 2014 Trial and Excel 2010 (non-trial)

    I'm using using MSSQServer 2014 Trial with Master Data Services installed. I have connected to the MDS server with the Excel 2010 MDS Add-In. 
    Models are not listed in the Add-In's Explorer browser. When I open the MDS server with IE, no models are listed either.
    I can view them locally on the server in IE.
    There are no remote DB access issues.
    Is it just that the 2014 Trial version does not supply this info remotely??

    Was a Model Permission issue. Fixed here: http://technet.microsoft.com/en-us/library/ff487055.aspx
    Would be good if the interface would show some kind of alert to that fact.

  • Master data attr./text/hier. loading

    According to best practice for Headcount and Personnel Actions( 0PA_C01), we have to do several process after infocube activation. e.g. active Query, query view, and loading transaction data.
    The question is when should we load master data attr./text/hier. ? Why does SAP do not put it on the best practice document?

    Hi,
    You should always load Master data before transactional data. This helps in loading transactional data faster as the SID tables are already populated once you load the master data and helps master data load performance. Make sure that the loaded master data is followed by an attribute change run always as if this step is not executed, the loaded master data will not take effect.
    Also in terms of reporting, if any users are accessing the queries and master data is missing at that point of time (might be loaded later), then the reports for those characteristics are going to show up as # (not assigned) and make no sense for reporting.
    If the referential integrity box is checked during transactional data load, the loads will fail if relevant master data is not loaded first.
    So, always load master data first, followed by attribute change run and then load the transactional data.
    Cheers,
    Kedar

  • Export Master Data to file with selected properties SAP BPC 10 NW

    Hello Gurus,
    Can you kindly help me with this. I am trying to export master data from one of my dimension using DM package "Export Master Data to File". The exported file should contain selected properties because the source dimension is having a lot of properties which are relevant for some usage but not required for the exported file.
    I used transformation file first to indicate the properties i wanted to be extracted under mapping, it extracted master data but it extracted all the properties. It's like it doesn't read from the transformation file.
    Does anyone encountered this scenario? Any other possible workaround to extract selected properties aside from opening the file and delete manually?
    Thanks as always,
    yajepe

    Hi Yajape
    I am unable to export the master data.
    Can you please tell me what is the mapping section you specified in transformation file.
    There is option in member selection screen called  member selection by properties through which you can select your properties.
    Regards
    Srikant

Maybe you are looking for

  • IPhone 4 From Hongkong?

    Hello, Currently am at Chengdu,CHINA,How can i get an iPhone 4 HongKong Version ,in chengdu many are offering iPHone 4 at high price..can any one help me regarding this...

  • New K8T NEO won't boot

    I'm a longtime computer fixer but new builder – New setup - first turn on, I get only the first MSI logo and am therefore unable to DEL into BIOS setup. The D-Bracket cycles through several combinations, but stops with red on 1 and green on 2, 3, and

  • Calling producer/consumer subvi using main vi for image grab acquire

    Hi everyone, I am not sure if I get the producer/consumer concept wrong. I tried to write a producer/consumer subvi that simulate continuously grab acquire image. When I call the subvi from the main while loop with a main vi, it seems to run into a i

  • Key words on IWeb site

    Hello, Hatman here. Is there some way possible to insert "key words" in you IWeb site so search engines can find out similar to the capability in Dreamweaver?

  • Macbook pro - no audio - no option for internal speakers in sound prefs.

    No audio will play from my 2011 macbook pro. It is fully updated. In sound preferences, there is no 'internal speakers' option to select in output, and no 'internal microphone' in input. No red light in headphone jack. Any solutions?