Two issues: activation of transfer rules and data load performance

hi,
I have two problems I face very often and would like to get some more info on that topics:
1. Transfer rules activation. I just finished transport my cubes, ETL etc. on productive system and start filling cubes with data. Very often during data load it occurs that transfer rules need to be activated even if I transport them active and (I think) did not do anything after transportation. Then I again create transfer rules transports on dev, transport changes on prod and have to execute data load again.
It is very annoying. What do you suggest to do with this problem? Activate all transfer rules again before executing process chain?
2. Differences between dev and prod systems in data load time.
On dev system (copy of production made about 8 months ago) I have checked how long it takes me to extract data from source system and it was about 0,5h for 50000 records but when I executed load on production it was 2h for 200000 records, so it was twice slower than dev!
I thought it will be at least so fast as dev system. What can influence on data load performance and how I can predict it?
Regards,
Andrzej

Aksik
1 How freequently this activation problem occurs. If it is one time replicate the datasource and activate thetransfer structure( But in general as you know activation of transfer structure should be done automatically after transport of the object)
2 One thing for difference of time is environmental as you know in production system so many jobs will run at the same time so obiously system performance will be slow compare to Dev System. In your case both the systems are performing equally. You said in dev system for 50000 records half an hour and in production 200000 records 2hrs so records are more in Production system and it took longer time. If it is really causing problem then you have to do some performance activities.
Hope this helps
Thnaks
Sat

Similar Messages

  • How to find unsued Transfer Rules and Data Sources for a Master Table...??

    How to find unsued Transfer Rules and Data Sources for a Master Table...?? My requirement is i need to delete those Transfer rules and Data Sources which are not in use

    Hi
    Go to manage of the Text or attirbute of the master data object, see what are being loaded daily from there and delete the remaining.
    Cheer
    Ans as expected, Please reward

  • How to debug a transfer rule during data load?

    I am conducting a flat file (excel sheet saved as a CSV file) data load.  The flat file contains a date field and the value is '12/18/1988'.  In transfer rule for this field, I use a function call to transfer this value to '19881218' which corresponds to BW DATS format, but the monitor of the InfoPackage shows red error:
    "Value '1981218' of characteristic 0DATE is not a number with 000008 spaces".
    Somehow, the last digit or character of the year 1988 was cut and the year grabbed is 198 other than 1988.  The function code is (see below in between two * lines):
    FUNCTION ZDM_CONVERT_DATE.
    ""Local Interface:
    *"  IMPORTING
    *"     REFERENCE(CHARDATE) TYPE  STRING
    *"  EXPORTING
    *"     REFERENCE(DATE) TYPE  D
    DATA:
    c_date(2) TYPE c,
    c_month(2) TYPE c,
    c_year(4) TYPE c,
    c_date_combined(8) TYPE c.
    data: text(10).
    text = chardate.
    search text for '/'.
    if sy-fdpos = 1.
      concatenate '0' text into text.
    endif.
    c_month = text(2).
    c_date = text+3(2).
    c_year = text+6(4).
    CONCATENATE c_year c_month c_date INTO c_date_combined.
    date = c_date_combined.
    ENDFUNCTION.
    Could experts here tell me what's wrong and also tell me on how to debug a transfer rule during data load?
    Thanks

    hey Bhanu/AHP,
    I find the reason.  Originally, I set the character length for the date InfoObject ZCHARDAT1 to 9, then I find the date field value (12/18/1988)length is 10.  Then I modified the InfoObject ZCHARDAT1 length from 9 to 10 and activated it already.  But when defining the transfer rule for this field, before the code screen, click the radio button "Selected Fields" and pick the filed /BIC/ZCHARDAT1, then continue to go to the transfer rule code screen, but find the declaration lines for the infoObject /BIC/ZCHARDAT1 is as following:
      InfoObject ZCHARDAT1: CHAR - 000009
        /BIC/ZCHARDAT1(000009) TYPE C,
    That means even if I've modified the length to 10 for the InfoObject and activated it, but somehow the transfer rule code screen always takes the old length 9.  Any idea to have it fixed to take the length 10 in the transfer rule code screen defination?
    Thanks

  • Unable to find Transfer Rules and Data Source for 0PERSON Info Source

    Hi,
    I'am working on SALES ANALYSIS Configuration with the help of SAP Best Practices.
    I'am trying to load Master Data Attributes/Texts/Hierarchies.
    I'am able to load for allmost all the InfoPackages which i created for particular DataSource.
    Now i need to load master data for 0PERSON InfoObject under that Person (Attributes) and Person (texts) are there,for this 2 Transfer Rules and the DataSource has to be defined by default as it is defined for other Attributes/Texts/Hierarchies.But i'am unable to find a DataSource and Transfer Rules to create a infopackage for PERSON Attributes and Texts.
    Please gudie me with a solution in order to get these Transfer Rules and DataSource for this Attributes and Texts.
    With regards,
    Pradeep.B

    Hi,
    Install the corresponding master data objects from Business content. Go to data source find out the needed datasource, replicate it. If you want you can migrate it also.
    If still you are not getting 0person masterdata object datasource & transformation structure, then you have to create the transformation, Infopackage.
    Hope if you will install the business content again, it will solve your problem.

  • QUERY PERFORMANCE AND DATA LOADING PERFORMANCE ISSUES

    WHAT ARE  QUERY PERFORMANCE ISSUES WE NEED TO TAKE CARE PLEASE EXPLAIN AND LET ME KNOW T CODES...PLZ URGENT
    WHAT ARE DATALOADING PERFORMANCE ISSUES  WE NEED TO TAKE CARE PLEASE EXPLAIN AND LET ME KNOW T CODES PLZ URGENT
    WILL REWARD FULL POINT S
    REGARDS
    GURU

    BW Back end
    Some Tips -
    1)Identify long-running extraction processes on the source system. Extraction processes are performed by several extraction jobs running on the source system. The run-time of these jobs affects the performance. Use transaction code SM37 — Background Processing Job Management — to analyze the run-times of these jobs. If the run-time of data collection jobs lasts for several hours, schedule these jobs to run more frequently. This way, less data is written into update tables for each run and extraction performance increases.
    2)Identify high run-times for ABAP code, especially for user exits. The quality of any custom ABAP programs used in data extraction affects the extraction performance. Use transaction code SE30 — ABAP/4 Run-time Analysis — and then run the analysis for the transaction code RSA3 — Extractor Checker. The system then records the activities of the extraction program so you can review them to identify time-consuming activities. Eliminate those long-running activities or substitute them with alternative program logic.
    3)Identify expensive SQL statements. If database run-time is high for extraction jobs, use transaction code ST05 — Performance Trace. On this screen, select ALEREMOTE user and then select SQL trace to record the SQL statements. Identify the time-consuming sections from the results. If the data-selection times are high on a particular SQL statement, index the DataSource tables to increase the performance of selection (see no. 6 below). While using ST05, make sure that no other extraction job is running with ALEREMOTE user.
    4)Balance loads by distributing processes onto different servers if possible. If your site uses more than one BW application server, distribute the extraction processes to different servers using transaction code SM59 — Maintain RFC Destination. Load balancing is possible only if the extraction program allows the option
    5)Set optimum parameters for data-packet size. Packet size affects the number of data requests to the database. Set the data-packet size to optimum values for an efficient data-extraction mechanism. To find the optimum value, start with a packet size in the range of 50,000 to 100,000 and gradually increase it. At some point, you will reach the threshold at which increasing packet size further does not provide any performance increase. To set the packet size, use transaction code SBIW — BW IMG Menu — on the source system. To set the data load parameters for flat-file uploads, use transaction code RSCUSTV6 in BW.
    6)Build indexes on DataSource tables based on selection criteria. Indexing DataSource tables improves the extraction performance, because it reduces the read times of those tables.
    7)Execute collection jobs in parallel. Like the Business Content extractors, generic extractors have a number of collection jobs to retrieve relevant data from DataSource tables. Scheduling these collection jobs to run in parallel reduces the total extraction time, and they can be scheduled via transaction code SM37 in the source system.
    8). Break up your data selections for InfoPackages and schedule the portions to run in parallel. This parallel upload mechanism sends different portions of the data to BW at the same time, and as a result the total upload time is reduced. You can schedule InfoPackages in the Administrator Workbench.
    You can upload data from a data target (InfoCube and ODS) to another data target within the BW system. While uploading, you can schedule more than one InfoPackage with different selection options in each one. For example, fiscal year or fiscal year period can be used as selection options. Avoid using parallel uploads for high volumes of data if hardware resources are constrained. Each InfoPacket uses one background process (if scheduled to run in the background) or dialog process (if scheduled to run online) of the application server, and too many processes could overwhelm a slow server.
    9). Building secondary indexes on the tables for the selection fields optimizes these tables for reading, reducing extraction time. If your selection fields are not key fields on the table, primary indexes are not much of a help when accessing data. In this case it is better to create secondary indexes with selection fields on the associated table using ABAP Dictionary to improve better selection performance.
    10)Analyze upload times to the PSA and identify long-running uploads. When you extract the data using PSA method, data is written into PSA tables in the BW system. If your data is on the order of tens of millions, consider partitioning these PSA tables for better performance, but pay attention to the partition sizes. Partitioning PSA tables improves data-load performance because it's faster to insert data into smaller database tables. Partitioning also provides increased performance for maintenance of PSA tables — for example, you can delete a portion of data faster. You can set the size of each partition in the PSA parameters screen, in transaction code SPRO or RSCUSTV6, so that BW creates a new partition automatically when a threshold value is reached.
    11)Debug any routines in the transfer and update rules and eliminate single selects from the routines. Using single selects in custom ABAP routines for selecting data from database tables reduces performance considerably. It is better to use buffers and array operations. When you use buffers or array operations, the system reads data from the database tables and stores it in the memory for manipulation, improving performance. If you do not use buffers or array operations, the whole reading process is performed on the database with many table accesses, and performance deteriorates. Also, extensive use of library transformations in the ABAP code reduces performance; since these transformations are not compiled in advance, they are carried out during run-time.
    12)Before uploading a high volume of transaction data into InfoCubes, activate the number-range buffer for dimension IDs. The number-range buffer is a parameter that identifies the number of sequential dimension IDs stored in the memory. If you increase the number range before high-volume data upload, you reduce the number of reads from the dimension tables and hence increase the upload performance. Do not forget to set the number-range values back to their original values after the upload. Use transaction code SNRO to maintain the number range buffer values for InfoCubes.
    13)Drop the indexes before uploading high-volume data into InfoCubes. Regenerate them after the upload. Indexes on InfoCubes are optimized for reading data from the InfoCubes. If the indexes exist during the upload, BW reads the indexes and tries to insert the records according to the indexes, resulting in poor upload performance. You can automate the dropping and regeneration of the indexes through InfoPackage scheduling. You can drop indexes in the Manage InfoCube screen in the Administrator Workbench.
    14)IDoc (intermediate document) archiving improves the extraction and loading performance and can be applied on both BW and R/3 systems. In addition to IDoc archiving, data archiving is available for InfoCubes and ODS objects.
    Hope it Helps
    Chetan
    @CP..

  • Relationship between ERPi metadata rules and Data load rules

    I am using version 11.1.1.3 to load EBS data to ERPi and then to FDM. I have created a Metadata rule where I have assigned a ledger from the Source Accounting Entities. I have also created a Data Load rule that references the same ledger as the metadata rule. I have about 50 ledgers that I have to integrate so I have added the source adapters that reference the Data Load rule.
    My question is... What is the relation between the Meatdata rule and the Data load rule with the ledger? If you can specify only one Ledger per metadata rule, then how does FDM know to use another metadata rule with another ledger attached to it? Thanks!!!

    Aksik
    1 How freequently this activation problem occurs. If it is one time replicate the datasource and activate thetransfer structure( But in general as you know activation of transfer structure should be done automatically after transport of the object)
    2 One thing for difference of time is environmental as you know in production system so many jobs will run at the same time so obiously system performance will be slow compare to Dev System. In your case both the systems are performing equally. You said in dev system for 50000 records half an hour and in production 200000 records 2hrs so records are more in Production system and it took longer time. If it is really causing problem then you have to do some performance activities.
    Hope this helps
    Thnaks
    Sat

  • Problems to activate transfer rules and infosource 2lis_02_itm and 2lis_02_

    hello,
    now I am working with the 0PUR_C01 cube of purchasing and I want to active the data flow before for it.
    I try to do it directly selecting data flow before in the activation settings of the cube, but I coludn´t
    after that, I try to do it activating the transfer rules and the infosource separately  but I get some errors, teh most important thing (because I have the same situation in other areas) is:
    The creation of the export DataSource failed
    Message no. RSBM035
    Diagnosis
    The creation of the DataSource &v1& was unsuccessful.
    System Response
    Whilst generating the export InfoSource, the metadata for a DataSource must be generated. An error arose when doing this.
    Procedure
    Consult the appropriate developer (function module: RSB1_OLTPSOURCE_GENERATE).
    I hope that some can help me with this issue,
    Thanks in advance and best regards,
    Niurka

    Thank you very much Chetal,
    I activated others Infocubes, but I have the same problem for Infocubes with ODS inside they data flow,
    I tried with some notes like: 538052 and I enter the client but not the user or the server, I proved the RFC Connection too.
    But the bug issue is the export data source creation that failled everytime, right now  I try to solve this and the log says some error about the syntax error in transfer routine for the infobject 0GN_R3_SSY this object exist as characteristic and his convertion routine is ok (because I verificated) and exist as part of the transfer rule but I don't know how to go to the routine creator for check this point.
    Thank you very much in advance,
    Niurka.

  • Programs for activation of transfer rules

    Can anybody give me the programs for activation of transfer ruls and update rules.

    Hi
    The programs are:
    RS_TRANSTRU_ACTIVATE_ALL(Transfer Structure activation)
    RSAU_UPDR_REACTIVATE_ALL(Update Rules activation)
    To activate all the source systems using the program RSAR_LOGICAL_SYSTEMS_ACTIVATE.
    Regards
    Shilpa

  • Failure while activating the transfer rules

    Hi,
    While activating the transfer rules, in that process when the system is trying to activating the transfer structure an error is occuring.
    Following is the error log.
    FREELISTS 004
    FREELIST GROUPS 01)
    PARTITION BY RANGE (PARTNO)
    PARTITION "/BIC/B00077630000000000002" VALUES LESS THAN
    (0000000002))
    ORA-00604: error occurred at recursive SQL level 1 ORA-01654: unable to extend index SYS.I_CON1 by 1
    DDL time(___1): .......167 milliseconds
    The SQL statement was not executed
    Table /BIC/B0007763000 (Error adjusting the database)
    Table /BIC/B0007763000 was not activated
    Obj_act >>(___1): ..00:00:01 hh:mm:ss
    Following is the log from System Log(SM21) for the same
    BY2 Database error 604 at EXE
    BY0 > ORA-00604: error occurred at recursive SQL level 1#ORA-01654
    BY0 > unable to extend index SYS.I_CON1 by 1065 in tablespace SYST
    BY2 Database error 604 at EXE
    Every time when I try to activate the transfer rules, it generates a new table "/BIC/B000xxxxxxx".
    Let me know if any more details required.
    Though there is a SAP Notes(513758) with Symptom which is same as our problem, its of not much use. It only says search for other SAP Notes with KeyWords from error log. We couldn't find a relavant SAP Notes sofar.
    Please guide us in resolving the issue.

    Hi Aswani,
    To further prevent this from occuring, you should assign a BASIS person to monitor the BW system using tcode <b>DB02</b> and watch out for critical objects prior loading process. That way, this error could have been prevented.
    You can also device a way to predict which table spaces are needed to adjusted on a scheduled basis based on your actual data growth.
    --Jkyle

  • Conversion exit error while activating the transfer rules

    Hello gurus,
    We are doing a function module extraction with 3.X data sources.
    On activating the transfer rules , i am getting a conversion exit error for "0fiscyear", but in the infoobject we don't have the conversion exit for it.
    Points will be awarded for any kind of response.
    Thanks and Regards,
    Anu.

    Hi,
    1. Please check in the transformation, its necessary to provide the fiscal variant mapping as well when ur mapping fiscal year object. as on the basis of that value fiscal period will be derived.
    2 I guess ur doing direct mapping (Time conversion) and there is no routine written for that.  if any u can check that as well for any error.
    Thanks
    dipika
    Edited by: Dipika Tyagi on Jun 25, 2008 7:39 AM

  • Unable to collect Update/Transfer rules Info & Data sources from BI content

    Hi,
    We are tyring to activate Infoprovider 0PY_CO2 from BI content by selecting the option - In data before. BI is unable to collect all corresponding objects such as Update rules, Info source, transfer rules and the data source for the selected Infoprovider. Only Info area and associated info objects are being collected for the installation.
    We are experiencing same problem whilst acitvating most of the info providers. However, in some cases, either of these objects get activated but not all the necessary objects that feed data to the Infoprovider (cube or multi provider).
    Also, Infopackages and Info object catalogs are not being collected for any objects activated/installed so fare from the BI content.
    We are trying to activate BW 3.5 content on the BI 7.0 environment. BI Content release is 703 and support pack level is 008.
    Kindly share any pointers to troubleshoot this probelm please.
    Thanks
    Venkat

    Shiv,
    For some Info providers, all associated info objects(including transfer & update rules and info source and data sources) are getting collected and got successfuly installed. But this is true for only 10% of the Info providers that we have installed so far.
    Remaining all the objects, we are unable to proceed further.
    Thanks
    Venkat.

  • Activation of transfer Rule not available:  Help Sought

    Hi All,
    I have encountered error in loading data due to
    dataSource not having the same status as the source system in the Business Information Warehouse.  There is a timestamp difference. The recommended solution in HELP was to replicate the datasource again and activate the transfer rules....  I have done the prescribed replication but have not been succesful in activating the transfer rules.  The activate button is disabled.  I tried to delete the transfer rule and assigned it again. But still the activate button is disabled?
    Thanks!  Hope you can help ASAP.
    Philips

    Hi,
    Replicate the datasource first. Then once you are done with that. Run the program RS_TRANSTRU_ACTIVATE_ALL in SE38. You will get a selection screen . Give the Source system name and the infosource name execute. This would activate your transfer rule. We use this in the production enviornment.

  • Activation of transfer rules...

    Hi Everyone,
    I can't seem to activate the transfer rule I defined...  Whatever could be wrong?  I've tried all of the regeneration of this and that to no avail.  Can anyone help me out please?  Thank you!
    Philips Manalaysay

    Hi,
    Re: Issue related to Transfer rules   
    Posted: Feb 26, 2007 8:15 AM    in response to: DK       Reply 
    U can activate the transfer structure by a program..
    execute this in SE38...
    RS_TRANSTRU_ACTIVATE_ALL..
    key in ur infosource and the source system name and execute
    TCODE -- SE16 AND EXECUTE above and then give the datasource name

  • Transfer Rules and Update Rules

    Hello,
    What is the exact difference between Transfer Rules and Update Rules ????
    NImisha

    Hi Nimesh,
       Transfer rules are the combination of communication structure and transfer structure here you are sassing the fields between transfer structure to communication structure..
       Updater rules are crating between data marts and info source, here you are creating the update rules based on communication structure.
       For more information go through the following link...
      http://help.sap.com/saphelp_nw04s/helpdata/en/fc/1251421705be30e10000000a155106/frameset.htm
    Regards,
    PRK
    Message was edited by: PRK

  • Document on XML date rules and date management

    Hi,
    Could you Please send a document on XML date rules and date management. I really in need of it. You may send it to [email protected]

    Hello
    Could you please email this to me:  [email protected]
    Thanks in advance
    Kevin

Maybe you are looking for

  • Can the portal language be specified in the URL

    can the portal read the language information from the launching URL? something like: The specified item was not found. The specified item was not found.

  • Open a external application in a concrete place in the screen

    I need to open an external application from my swing app, but the application window should be opened in a concrete place in the screen, such as using setBounds in a JInternalFrame. Runtime methods just run the application in a new thread, but I can'

  • Specials characters in subject mail

    Hello, We created a procedure in Oracle to send html mails. All is allright in the body, the only problem is in the subject : we can't have characters like "é è à"... Our database is Oracle 9i.. so we don't have the UTL_ENCODE.MIMEHEADER_DECODE funct

  • Time Machine filled 300GB HD Now what?

    Hi ! I am new to Leopard....! My Time Machine is full after backing up since March 11th. I have deleted the only choices I seemed to have ... systems, library and finally Backup. Still I only have 4GB left on a 300GB hard drive. I don't know how to e

  • Access certificate from a DoD CAC card with Sun Java 1.4.2 Plug-in

    We are using CAC cards with ActivCard Software. We are using Sun Java JVM 1.4.2 with Internet Explorer. The problem is that Sun Java does not use the certificate in the browser certificate store for HTTPS client authentication. The solution from Sun