Transport  Steps prob

Hi ,
i am transporting BCT cube from DEV to Q system
i fallow this steps
-->  i am using only neccary option in transport connection and collect
  based on the Flow of IC tree.
  1 . drag IC -- it co,llects all Info areas and Info objects
  2. drag IS -  it collects App comp's and Comm Structure
  3. drag DS- it collects only one data sources under IC tree.
  4. Drag Update rules --it collects all routines , IC and Info sources.
  5. Drag Transfer rules -- it collects Sources system , IS and info objects.
and all are store in one request .
my doubt is how can collect Master data sources ?
and tell me i fallow correct order or not ?
i assign points..
Regards,
PSR

Hi PSR,
Why don't you collect all at once by selecting as "Indata Flow Before" in Transport connection, it will collect all objects related to that IC.
Check the required one only.
Coming to Master data Sources you can't transport them.
you have to create them seperately in other servers as QA nd Production.
As the way available in Dev.
Hope it helps .
Reg
Pra

Similar Messages

  • Tp executes 10 second sleep function for every transport step

    I am currently working on a problem related to performance of transports. In all of the environments i work with (about 50 in total), be it DEV,ACC,QAS,SBX or PRD, i have noticed considerable delays when tp is performing an import. This delay is occurring for every step that has to kick off an RDD-job. The jobs themselves finish very quickly, within 1 or 2 seconds.  I very much convinced tp is waiting for 10 seconds exactly after kicking off a job in SAP via a trigger.  After tracing tp with flags '-t' and 'tf=' i noticed in the trace references to this 10 second wait. I did debug one of the latest tp versions on a Windows platform and found the code that does this. A PUSH of the number 10 and a CALL to a sleep routine. When you execute this code and time it it takes 10 seconds before tp comes back and then start checking the successful completion of the job it kicked off. I have changed the value of 10 to 1 with a hex editor and now the delay is only 1 second and my transports are still running fine. The duration of a very simple transport with 4 steps inside a batch job in SAP dropped from 45 to 5 seconds.
    Perhaps you think, why bother at all, but when you consider that these steps are executed multiple times per transport (sometimes even 10 times or more) and that frequently bundles are imported where several thousands of transports go in you understand why i am persuing a better solution.
    There may be a valid reason why tp is waiting, but then the mechanism has to improve.
    My question is, has somebody also noticed that tp is unnecessarily delaying the imports and what have you done to investigate ?
    Can somebody confirm what i am claiming ?
    Thanks in advance, Peter van der Woude

    For those of you who are interested in speeding up their imports without having to use 'import all' version 340.16.25 will be released shortly. In this version the sleeps are reduced from 10 to 1 second per transport step. Average transport times should drop between 40 to 50 seconds. Depending on the contents the gain could also be only 20 seconds.

  • Implement transport step by step.

    Hello Gurus,
             please instruct me whole detail to implement transport step by step.
    Many thanks

    Hi,
       When you are collecting the objects the most important thing to consider is the sequence of the Transports. You always have to send the objects in the same order in which you have created them. Sequence will be as below:
    1. Data Source in Source system
    2. Infoareas
    3. Infoobject Catalogues & Infoobjects
    4. Infosource & Transfer rules
    5.Data Targets
    6. Update Rules/Transformatioms
    7. Infopackages/DTPs
    8. Multipprovders/Infosets
    9. Reporting Elements
        Always the requests should be sent in the same sequence if different types of objects are collected in different requests. If the object list is small you can collect everything into a singel request.
    Thanks,
    Praveen

  • Business Content instalation & Transportation steps

    Hi,
    iam a new in Sap bi implementation project.so iam gettiong somany doubts abot Business Content instalation & Transportation
    can you help me above concepts step by step,
    Thanks,
    chal

    Please search the forum before posting a thread
    Please do not post duplicate threads
    Business Content instalation & Transportation steps

  • Master data transport steps

    Hi expert,
    What will be the steps to transport master data in BI 7? I have transported relevant R/3 data source and then replicated data source in BW. Then I start collecting BW info objects for the master data. When collecting BW info objects for transport, do I need include data source, transfer rule and info pacakges? If so, why do I need to collect data source again in BW system as it has already been replicated from the source system?
    Thank in advance for any advise.
    Sharon

    Hi,
    The procedure for loading master data in 7.0 is :
    1. Create / Activate your data source for the info object you want load - for example data source for attributes, text ot hierarchy - SAP R/3
    2. Replicate data source in SAP BW
    3. Create Data Transformation ( similar to Update and Transfer rules) between Data Source and Info Object
    4. Create DTP ( data transfer process) to maintain the data flow . Here you have to specify that data comes from Data Source and goes to the Info Object.
    5. Create an Info Package on the Data Source and load data - !!! have in mind that this will load data only into PSA table
    6. Final step - you have to start your DTP - this will load data from PSA to the Info Object.
    For the majority of Master Data, there will be datasources in Business Content.  need to install Business Content . This will install all datasources that you require - Customer, Material etc. plus all standard config - DSO, Cubes etc.
    http://help.sap.com/saphelp_nw2004s/helpdata/en/80/1a66d5e07211d2acb80000e829fbfe/frameset.htm
    Pls Check the following threads for DTP
    DTP EXPLANATION
    Delta DTP request in Yellow color
    http://help.sap.com/saphelp_nw2004s/helpdata/en/42/fa50e40f501a77e10000000a422035/frameset.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/42/fbd598481e1a61e10000000a422035/frameset.htm
    Hope this helps,
    Regards
    CSM Reddy

  • Transports steps after Homogeneous Copy

    I'm planning to do a system copy in a few weeks time.  I have done dummy runs and all is fine, but i'm unsure about the steps that are required for the transpors after the live copy.
    The installation notes are very vague on this front.  Is it just a case of running se06, or do I need to set up the transports domain again etc??
    Thanks
    Ian

    You've probably already thought of this, but just as a reminder:
    1) you'll have to log onto client 000 and request to be added back to the Transport domain, then activate the request on the Domain Controller (tcode stms)
    2)Applied transport request level may be different between your QA & Prod. I usually down load the E070 table of QA before the refresh and compare it to the a download after the restore. Then manually add the transports to bring QA back to the transport level is was before the refresh. (May not be a concern in your landscape)

  • BW APO Transport steps (also please look into the error).

    Hi all,
          We have created data marts from BID to APOD and the loads are fine, I am ready with the trasports from APO dev to APO quality. I captured the replication (datasources), Transfer Structure, comm structure, update rules and infopackage, process chain .. everything is on APODev side..When i trasported orderly,we are getting this error.
    Object 'DM' (APCO) of type 'Application' is not available in version 'A'
    Message no. RSO252
    You wanted to generate an object with the name 'DM' (in transport request APCO) of type 'Application' (TLOGO). This is, however, not available in the BW Repository database. It does not exist in the requested version A. If the version is 'D' then it is possible that an error arose during the delivery or installation. If the version is 'A' then the Object was either not created or not activated.
    Please give me the resolution  with the above error, please tell me  what all i need to capture when i trasnport the datamrts.Do i need to  cature anything on BID also...or only in APO DEv.
    Thanks
    arya

    Hi Sam,
       They are on different appliction servers,  I have done the steps in APO QA some time back..i replicated the data source on APO QA side.. and i tried to move the transports , still i get the error..
    I am confused at one point and need some help in clarification. I have ODS on BW side and Cube on APO side, so when i do the generate export data source on BW side. should i send the ODS activation from BID to BIQ as well, My point here is should i transport anything on BW side (BID to BIQ) apart from APO DEV to APO QA,
    i have all transports ready on APO Dev side and i am getting the error,
    Please let me know if i can give more details.. thanks for your answers

  • User decision step prob

    Hi,
    In my workflow user decision step have 2 approvers.
    if one of the approver opening the work item ,remaining approvers relevent work item(User decision) deleted from sap inbox.
    i want to keep all approvers work items  alive upto getting decision from any one.
    Please suggest me in  this
    Thanks and Regards,
    Ragavendran K.

    Hi,
    It is  possible by doing the following
    create a multiline element for the approvers.
    you use this element in a simple subflow where you pass on the mutliine element in the Miscellaneous  tab.
    Now for every record in the container 1 workflow is started.
    Your workflow should have a fork 1:2
    in one branche the user decision and after the user decision a create event step.
    And in the other branch a wait for event step.
    Now only when a UD is executed an event gets triggered,m which should terminate all of the workflows.
    Kind regards, Rob Dielemans

  • Transport control program tp ended with error code 0203

    when i import queue and transport queue from dev to prd ,error show:
    Transport control program tp ended with error code 0203
    Message no. XT200
    Diagnosis
    An error occurred when executing a tp command.
      Command: IMPORT DEVK900298 P01 clientP01 U0 pf=/usr/sap/tra
      Return code: 0203
      Error text: wrong syntax in tp call
      Request: DEVK900298
    System Response
    The function terminates.
    Procedure
    Correct the error and execute the command again if necessary.

    Dear Allen,
    The transport control program tp collects all return codes that occur when a tp command is executed. From all of these single return codes, a combined return code is calculated that tp outputs as a result at the end the command. During an import, for example, all return codes for individual import steps are handled. If a special return code does not occur, then the combined return code at the end is the maximum of individual return codes that occur. In addition to the value of the return code, tp also gives a short description of this return code. You can display this short description using the command explainrc.
    Normally, all return codes that occur originate from individual transport steps. Since these individual return codes can only have the following values 0, 4, 6, 8, 12, 13, 14, 16, the combined return code also has one of these values. However, if a general error occurs, a single return code over 200 is generated. For example, this could arise if tp cannot establish a connection to the database. The maximum return code for all single transport steps is output in a separate message before specifying the combined returned code. In this case, the combined return code is always the return code of the general error message.
    <b>tp Commands:</b>
    tp addtobuffer <request> <sapsid>
    As of Release 4.6A, an explicit addtobuffer command leaves the position of the entry unchanged, if the relevant transport request is already in the buffer. Only the status of the entry is changed:
    Unconditional modes appearing in the command line are added to the unconditional modes in the buffer.
    All the steps of the entry are re-initialized.
    If the entry found is not an initial entry, the import step for the command file is set to *4000 .
    Before you call this command and its options, consider the effects of changing the import sequence.
    tp cleanbuffer <sapsid>
    This deletes successfully imported change requests from the list of requests that are marked for import into the SAP System <SAPSID>. This function is contained in the commands tp import all <sapsid> and tp put <sapsid> .
    tp delfrombuffer <request> <sapsid>
    If the specified request is marked for import into the specified SAP System, this flag is deleted from the import list.
    Before you call this command, consider the effects of changing the import sequence.
    As of Release 4.6A: If a transport request was only partially imported, you can delete this type of transport request from the buffer. However, if a deleted transport request is placed in the buffer (without having been initially or completely imported), the attributes of the old entry are reactivated. This means the return codes, the unconditional modes and tags of the deleted entry appear in the new entry.
    tp delstopmark <sapsid>
    If the list of requests marked for import into the specified SAP System contains a STOP mark, the mark is deleted.
    tp export <request>
    This exports the complete request from the source system. This command starts the export of a request from the operating system level. Only use this command in exceptional cases. The SAP System uses the command expwbo to release requests from CTS transactions.
    tp reformatbuffer <sapsid> [options]
    This automatically reformats the entire buffer.
    As of Release 4.5A, there are 2 different formats for buffer entries. (As of Release 4.5A, tp can work with both formats, and with a puffer containing heterogeneous entries. As of Release 4.6A, tp can work with 3 formats.)
    However, in certain cases entries with the old formatting have to be reformatted during the normal run of tp . In this case, the entire buffer is automatically reformatted.
    If you use the options downtorelease and Iagree for command reformatbuffer , you can format the buffer so that its formatting is compatible with an older release:
    Option downtorelease <Release>
    The command reformatbuffer with this option formats the buffer so that the formatting corresponds to the one used in an old release. This is necessary if tp writes buffer entries to the buffer of another system, but uses a buffer format that the transport control programs does not understand in this system.
    Since older formats cannot store the same amount of information as the newer ones, some information contained in the original buffer may not be stored in the unformatted buffer. This may lead to data loss.
    To avoid data loss, the original buffer is stored. In addition, you have the following functions:
    Displaying the unformatted buffer (using showbuffer <SID> )
    Displaying the original buffer (using showbuffer buffer=<SID>SAV )
    Accepting the unformatted buffer (using reformatbuffer and the option Iagree yes )
    Resetting the changes (using reformatpuffer and the option Iagree no )
    If you call reformatpuffer using downtorelease , tp displays in detail if information was lost by reformatting, and what that information was. tp also names the commands that is has to execute to accept or reject the changes.
    If you use the option downtorelease , you must also specify the option Iagree with yes or no . You cannot use the transport system until you decide to accept or reject the changes.
    Option Iagree
    Here are the following values for this option:
    yes ( true )
    Accept the change made by tp reformatbuffer <SID> downtorelease <rel>
    In this case, tp deleted the buffer semaphore file that was lifted for security reasons at the first call of reformatbuffer
    no ( false )
    Reject the change made by tp reformatbuffer <SID> downtorelease <rel> In this case, tp copies the backup of the original buffer back and deletes the buffer semaphore file that was lifted for security reasons at the first call of reformatbuffer
    tp setstopmark <sapsid>
    A special STOP mark is added to the list of requests registered for import into the specified SAP System. The commands tp import all <sapsid> and tp put <sapsid> only process the requests in front of this stop mark.
    If there is no stop mark, it is automatically generated by the commands
    tp import all <sapsid> and tp put <sapsid> .
    When all the requests in front of the stop mark have been imported successfully, at the end of all the calls the related marks and the mark itself are deleted.
    A STOP mark can only be placed at the end of the change requests that have already been marked. In addition, it can only be set if no other mark of the same kind already exists.
    tp showparams <sapsid>
    This tp function displays which values the individual parameters of the transport profile have for the current SAP System.
    tp verse <request>
    This creates versions of the objects in the specified request. This transport step is contained in command export .
    As of Release 4.5A, more actions that prepare the export are performed in this step.
    Thanks and Regards,
    Naveen.

  • Transport

    Hi All,
    I have completed of creating objects and testing in development for a project.
    I have created infoobjects, infosources, ods's and cubes and now i have to send them to quality system.
    i have a task before sending them.It is to organize the objects in a specific packages and requests.
    can we change the package and request no for an object?
    can we keep all the infoobjects in one request and in one package and similarly to infosources , ods's and cubes?
    if yes, how?
    is there any procedure to follow when transporting them into the quality?like sequence etc...
    and what is the procedure to follow?
    Kindly explain in detail of this procedure..
    Thanks and Regards,
    Pooja

    hi,
         select any one of the object example infocube, double click on the infocubeselect extra's menuthen select write transport request--there you can change the package name and save under it.
    you can do it for all the object's like this.
    standard transportation steps follows
    *RSA1-> Transport Connection
    *Select SAP Transport Folder.
    *Choose your respective objects(It can be Infobjects, infoproviders or infosources)
    *Drag and drop the objects which need to be transported to the right hand side of
    collected objects
    *click on truck icon(Transport Objects)
    *Specify the package name
    *save
    *A new request would be created(if not create a new request )
    *goto se09
    *search for the request number that is created
    *release
    In general sequence ...
    1. Info object
    2. Infosource,transfer rule
    3. Data Target,info provider
    4. Update rules
    5. queries
    6. Webtemplates,workbooks
    7. Roles
    hope this help you
    regards
    harikrishna N

  • Transport classes and programs to another server

    Except for transport routing and charm copy function, do we have other ways to do so?
    for example copy or so forth.
    best regards,
    Blake Le

    Hi Blake
    SAP R/3 Correction and Transport System
    Operating system level files in the transport process:
    The SAP C program TP, requires a special file structure for the transport process. The file system is operating system dependent. TP uses a transport directory or file system, which is called /usr/sap/trans.
    The /usr/sap/trans file system is generally NFS mounted form the development system to other systems unless a system is defined as a single system in the CTS pipeline. All the sub directories should have <SID>adm as the owner and sapsys as the group; and proper read, write and execute access should be given to owner and the group. The TP imports are always performed by <SID>adm.
    The following are the subdirectories in /usr/sap/trans:
    /data
    /cofiles
    /bin
    /log
    /actlog
    /buffer
    /sapnames
    /tmp
    /usr/sap/trans/data: holds the data of transport objects after they are released . The example of a data file is R904073.DEV. The extension DEV means the data file was released from the DEV or development system.
    /usr/sap/trans/cofiles: The cofiles directory holds the command files for all change requests. These files are like a command or control files used to import the data files. The common directory for CTS system is /usr/sap/trans. After a change request is released from the source system , the data is exported immediately to the file system of the operating system. The SAP transport utility TP uses the cofile to transport a data file. The example of a file in cofiles directory is K904073.DEV.
    /usr/sap/trans/bin: holds the most important file TPPARAM in the CTS system. TPPARAM file has all the information about the CTS systems in the CTS pipeline. TPPARAM file is the parameter file for the transport program TP and it is the common file for all the systems in the CTS pipeline. As you know already that /usr/sap/trans should be NFS mounted to all the systems in a CTS pipeline, TP program has access to the TPPARAM file from all the systems. The following is an example of typical TPPARAM file for five SAP systems in the CTS pipeline:
    #@(#) TPPARAM.sap 20.6 SAP 95/03/28
    Template of TPPARAM for UNIX #
    First we specify global values for some parameters, #
    later the system specific incarnation of special parameters #
    global Parameters #
    transdir = /usr/sap/trans/
    dbname = $(system)
    alllog = ALOG$(syear)$(yweek)
    syslog = SLOG$(syear)$(yweek).$(system)
    System spezific Parameters #
    Beispiel T11 #
    DEV/dbname = DEV
    DEV/dbhost = sap9f
    DEV/r3transpath = /usr/sap/DEV/SYS/exe/run/R3trans
    QAS/dbname = QAS
    QAS/dbhost = sap8f
    QAS/r3transpath = /usr/sap/QAS/SYS/exe/run/R3trans
    TRN/dbname = TRN
    TRN/dbhost = sap17
    TRN/r3transpath = /usr/sap/TRN/SYS/exe/run/R3trans
    PRE/dbname = PRE
    PRE/dbhost = sap19f
    PRE/r3transpath = /usr/sap/PRE/SYS/exe/run/R3trans
    PRD/dbname = PRD
    PRD/dbhost = sap18f
    PRD/r3transpath = /usr/sap/PRD/SYS/exe/run/R3trans
    /usr/sap/trans/log: holds the entire log files, trace files and statistics for the CTS system. When the user goes to SE09 (workbench organizer) or SE10 (customizing organizer) transaction and opens the log for a transport, the log file for that transport will be read from /usr/sap/trans/log directory. Each change request should have a log file. Examples of log files are DEVG904073.QAS, DEVI904073.QAS and DEVV904073.QAS. The name of a log file consists of the names of the change request, the executed step, and the system in which the step was executed:
    <source system><action><6 digits>.<target system>
    Now we can analyze the above example DEVG904073. QAS. The <source system> = DEV, <action> = G or report and screen generation, <6 digits> = 904073 (these six digits numbers are exactly the same number as the six digits of the transport) and the <target system> = QAS
    Possible values for <action> are:
    A: Dictionary activation
    D: Import of application-defined objects
    E: R3trans export
    G: Report and screen generation
    H: R3trans dictionary import
    I: R3trans main import
    L: R3trans import of the command files
    M: Activation of the enqueue modules
    P: Test import
    R: Execution of reports after put (XPRA)
    T: R3trans import of table entries
    V: Set version flag
    X: Export of application-defined objects.
    /usr/sap/trans/actlog: holds action log files. The example of an action file is DEVZ902690.DEV. The following are the contents of the file:
    1 ETK220 u201C==================================================u201D u201C=================
    =============================
    1 ETK191 u201C04/30/1998u2033 Action log for request/task: u201CDEVK902690u2033
    1 ETK220 u201C==================================================u201D u201C=================
    =============================
    1 ETK185 u201C04/30/1998 18:02:32u2033 u201CMOHASX01u2033 has reincluded the request/task
    4 EPU120 Timeu2026 u201C18:02:32u2033 Run timeu2026 u201C00:00:00u2033
    1 ETK193 u201C04/30/1998 18:02:33u2033 u201CMOHASX01u2033 owner, linked by u201CMOHASX01u2033 to u201CDEVK902691u2033
    4 EPU120 Timeu2026 u201C18:02:33u2033 Run timeu2026 u201C00:00:00u2033
    1 ETK190 u201C05/04/1998 11:02:40u2033 u201CMOHASX01u2033 has locked and released the request/task
    1 ETK194 u201C05/04/1998 11:02:40u2033 **************** End of log *******************
    4 EPU120 Timeu2026 u201C11:02:40u2033 Run timeu2026 u201C00:00:09u2033
    ~
    ~u201DDEVZ902690.DEVu201D 10 lines, 783 characters
    /usr/sap/trans/buffer: transport buffer of the target systems; contains control information on which requests are to be imported into which systems and in what order the imports must occur. The /usr/sap/trans/buffer will have a directory for each system in the CTS pipeline. For example the buffer file for DEV system is /usr/sap/trans/buffer/DEV.
    /usr/sap/trans/sapnames: holds information pertaining to transport requests for each system user. There are files for each user who released change requests from the system.
    /usr/sap/trans/tmp: holds information about temporary data and log files. While the transport is occurring the Basis administrator can find a file that is related to the transport in the tmp directory; that file shows the exact status if the transport (What objects are being imported at that time).
    Important SAP delivery class and table types and tables in the CTS process:
    Delivery class
    The delivery class defines who (i.e. the SAP system itself or the customer) is responsible for maintaining the table contents. In addition the delivery class controls how the table behaves in a client copy and an upgrade. For example when you select a SAP defined profiles to perform a client copy, certain tables are selected according to their delivery class. DD02L table can show what delevery class a table belongs to.
    The following delivery classes exist:
    A: Application table.
    C: Customizing table, maintenance by customer only.
    L: Table for storing temporary data.
    G: Customizing table, entries protected against overwriting.
    E: Control table.
    S: System table, maintenance only by SAP.
    W: System table, contents can be transported via own TR objects.
    Table type
    The table type defines whether a physical table exists for the logical table description defined in the ABAP/4 Dictionary and how the table is stored on the database.
    The following are different table types in SAP:
    Transparent Tables
    There is a physical table on the database for each transparent table. The names of the physical table and the logical table definition in the ABAP/4 Dictionary are same. For every transparent table in SAP, there is a table in database. The business and application data are stored in transparent tables.
    Structure
    No data records exist on the database for a structure. Structures are used for the interface definition between programs or between screens and programs.
    Append Structure
    An Append structure defines a subset of fields which belong to another table or structure but which are treated as a separate object in the correction management. Append structures are used to support modifications.
    The following table types are used for internal purposes, for example to store control data or for continuous texts:
    Pooled table
    Pooled tables can be used to store control data (e.g. screen sequences, program parameters or temporary data). Several pooled tables can be combined to form a table pool. The table pool corresponds to a physical table on the database in which all the records of the allocated-pooled tables are stored.
    Cluster table
    Cluster tables contain continuous text, for example documentation. Several cluster tables can be combined to form a table cluster. Several logical lines of different tables are combined to form a physical record in this table type. This permits object-by-object storage or object-by-object access. In order to combine tables in clusters, at least part of the keys must agree. Several cluster tables are stored in one corresponding table on the database.
    Tables in CTS process:
    TRBAT and TRJOB:
    TRJOB and TRBAT are the major tables in the CTS process. After TP program has sent the event to the r3 system, RDDIMPDP checks table TRBAT in the target system to find out if there is an action to be performed. Mass activation, distribution, or table conversions are the examples of actions. If there is action to be performed, RDDIMPDP starts the appropriate program in the background task. RDDIMPDP then reschedules itself.
    By checking table TRJOB, RDDIMPDP automatically recognizes if a previous step was aborted, and restarts this step. For each transport request , TP program inserts an entry into table TRBAT. If the return code 9999 in this table then the step is waiting to be performed. Return code 8888 indicates that the step is active and currently being processed. A return code of 12 or less indicates that the step is finished. In addition, TP inserts a header entry to let the RDDIMPDP program know to start processing. The column return code will therefore contain a B for begin. When RDDIMPDP is started, it sets the header entry to R(un), and starts the required program. When all the necessary actions are performed for all the transport requests, the column return code contains all the return codes received, and the column TIMESTAMP contains the finishing time. The header entry is set to F(inished). TP monitors the entries in TRBAT and TRJOB tables. When the header entry in TRBAT is set to finished. The entry in TRJOB is deleted.
    Transport Tables SE06
    TDEVC - Development classes
    TASYS - Details of the delivery. Systems in the group that should automatically receive requests, have to be specified in table TASYS.
    TSYST - The transport layers will be assigned to the integration systems. ( Define all systems)
    TWSYS - Consolidation routes ( define consolidation path)
    DEVL - Transport layers are defined here
    In u201CConfiguring the CTS systemu201D section, We will learn more about the transport tables in SE06 transaction
    Programs in the CTS process:
    In the CTS table section we learned about the RDDIMPDP program. RDDIMPDP program needs to be scheduled in all the clients in an instance. It is recommended to schedule the RDDIMPDP as event driven.
    RDDPUTPP and RDDNEWPP programs can be used to schedule RDDIMPDP program in the background.
    The ABAP/4 programs that RDDIMPDP starts are determined by the transport step to be executed that is entered in the function field of table TRBAT.
    Function Job Name Description of transport Steps
    J RDDMASGL Activation of ABAP/4 dictionary objects
    M RDDMASGL Activation of match codes and lock objects
    S RDDDISOL Analysis of database objects to be converted
    N RDDGENOL Conversion of database objects
    Y RDDGENOL Conversion of matchcode tables
    X RDDDICOL Export of AD0 objects
    D RDDDIC1L Import of AD0 objects
    E RDDVERSE Version management update during export
    V RDDVERSL Version management update during import
    R RDDEXECL Execution of programs for post - import processing
    G RDDDIC3L Generation of ABAP/4 programs and screens
    Version Management:
    One of the important features of Workbench Organizer is Version Management. This feature works for all the development objects. Using the version management feature the users can compare and retrieve previous versions of objects.
    Version management provides for comparisons, restore of previous versions, documentation of changes and assistance in the adjustment of data after upgrading to a new release. With the release of a change request, version maintenance is automatically recorded for each object. If an object in the system has been changed N times, it will have N delta versions and one active version. To display version management, for ABAPs use transaction SE38 and for tables, domains and data elements use SE11. The path to follow is Utilities -> Display version. Using version management the users can view existing version for previously created ABAP code, make changes to the code, compare code versions and restore original version of the code. Now the users will be restore previous versions without cut and paste steps of the past.
    TP and R3trans program:
    The basis administrator uses TP program to transport SAP objects from one system to another. TP is a C program delivered by SAP that runs independently of the R/3 system. TP program uses the appropriate files located in a common transport directory /usr/sap/trans. TP starts C programs, ABAP/4 programs and special operating system commands to its job. R3trans is one of the most important utility program called by TP. Before using the TP program, the basis administrator needs to make sure that the CTS system is setup properly and the right version of TP is running in the system. The TP program is located in the run time directory /usr/sap/<SID>/SYS/exe/run directory. It is automatically copied in the install process. A global parameter file TPPARAM that contains the databases of the different target systems and other information for the transport process controls TP. The global parameter file determines which R3trans is used for each system. If the parameter r3transpath is not defined properly then no export and import can be done. The basis administrator should make sure that the default value u201Cr3transpathu201D is properly defined. Later in this chapter we will learn more about TP and R3trans; also we are going to see how they are used.
    Configuring the TPPARAM file:
    Each time TP is started, it must know the location of the global parameter file. As we have seen before TPPARAM file should be in directory /usr/sap/trans/bin. The parameters in TPPARAM can either global (valid for each and every system in the cts pipeline) or local to one system. Th parameters are either operating system dependant (these parameters preceded by a keyword corresponding to the specific operating system) or database dependant (contain a keyword corresponding to a specific database system).
    The global parameter file provides variables that can be used for defining parameters. The variables can be defined in format: $(xyz). The brackets can be substituted with the u201C\u201D-character if required.
    The following pre-defined variables are available for the global parameter file:
    $(cpu1): The CPU name can be sun or as4 for example. In heterogeneous networks this variable is very important.
    $(cpu2): Acronym for the name of the operating system. The example for this variable can be
    hp-ux, or sunos . This is an operating system specific variable.
    $(dname): Used for the day of the week (SUN,MON,u2026.).
    $(mday): Used for the day of the current month (01-31).
    $(mname): Used for the name of the month (JANu2026DEC).
    $(mon): Used for the Month (01-12).
    $(system): R/3 System name.
    $(wday): Day of the week (00-06, Sunday=00, Monday=01, Tuesday=02 and so on).
    $(yday): Day of the current year (001-366). Using the number any day of the year can be chosen.
    $(year): Year (Example:1998 or 1999).
    $(syear): Short form of the year (two positions).
    $(yweek): Calendar week (00-53). The first week begins with the first Sunday of the year.
    For the database connection:
    The transport environment also needs parameters to connect to the R/3 System database. As we know already the every instance in the R/3 CTS pipeline has its own database, therefore specific parameters should be defined for each database system. From dbtype parameter of RSPARAM file, TP program identifies the database system.
    The two parameters u201Cdbnameu201D and u201Cdbhostu201D are required for ORACLE databases.
    DBHOST: is the name of the computer on which the database processes execute. TCP/IP name of the host if NT is being used.
    DBNAME: is the name of the database instance.
    As of Release 3.0E, two new parameters have been introduced.
    DBLOGICALNAME: The default value is $(system). The logical name that was used to install the database.
    DBCONFPATH: The default value is $(transdir).
    The parameters u201Cdbnameu201D and u201Cdbhostu201D are also used for INFORMIX databases in an installation:
    DBHOST: Same as Oracle.
    DBNAME: Name of the database instance, uppercase and lowercase are distinguished here.
    INFORMIXDIR : u201C/informix/<SAPSID>u201D is the default value. Defines the directory namewhere the database software can be found.
    INFORMIXSQLHOSTS: u201C$(informixdir)/etc/sqlhosts[.tli|.soc]u201Cis default value under Unix. The name of the SQLhosts file with its complete path is defined with this parameter.
    INFORMIX_SERVER: u201C$(dbhost)$(dbname)shmu201D is the default value. The name of the database server may be specified for a local connect.
    INFORMIX_SERVERALIAS: u201C$(dbhost)$(dbname)tcpu201Dis the default vlue. The name of the database server can be specified for a remote connect.
    For Microsoft SQL Server database the two parameters u201Cdbnameu201D and u201Cdbhostu201D are also required. DBHOST: The TCP/IP name of the host on which the database is running.
    DBNAME: The database instance name.
    For DB2 in AS/400 only u201Cdbhostu201D is required.
    DBHOST: System name of the host on which the database is running.
    If theu201DOptiConnectu201D is used, the following line should be specified:
    OPTICONNECT 1
    For DB2/ AIX
    The two parameters u201Cdbnameu201D and u201Cdbhostu201D are required
    DBHOST: The host on which the database processes are running. It is the TCP/IP name of the host for Windows NT (As we have seen in the earlier examples).
    DBNAME: Database instance name.
    The DB2 for AIX Client Application Enabler Software must also be installed on the host on which tp is running.
    ALLLOG: u201CALOGu201D $(syear) $(yweek)u201Dis the default value. This variable can be used in TPPARAM file to specify the name of a file in which tp stores information about every transport step carried out for a change request anywhere in the transport process. The file always resides in the log directory.
    SYSLOG: u201CSLOG $(syear) $(yweek).$(system)u201D is the default value. This variable can be used to name a file in which tp stores information about the progress of import actions in a certain R/3 System. The file does not store information for any particular change request. The file always resides in the log directory.
    tp_VERSION: Zero is the default value. If this parameter is set to not equal to zero, a lower version of tp may not work with this TPPARAM file. If the default value (zero) is set, the parameter has no affect.
    STOPONERROR: (Numeric value) The default value is 9. When STOPONERROR is set to zero, tp is never stopped in the middle of an u201Cimportu201D or u201Cputu201D call. When STOPONERROR is set to a value greater than zero, tp stops as soon as a change request generates a return code that is equal to or greater than this value (The numeric value of the STOPONERROR parameter is stored in the variable BADRC). Change requests, which still have to be processed for the current step, are first completed. A u201CSYNCMARKu201D in the buffer of the R/3 System involved, sets a limit here. tp divides the value of this parameter between two internal variables. STOPONERROR itself is treated as a boolean variable that determines whether tp should be stopped, if the return code is too high.
    REPEATONERROR (Numeric value too): The default value is 9. The REPEATONERROR parameter is similar to STOPONERROR. The difference is, REPEATONERROR specifies the return code up to which a change request is considered to be successfully processed. Return codes less than REPEATONERROR are accepted as u201Cin Orderu201D. Change requests that were not processed successfully stay in the buffer.
    NEW_SAPNAMES: Default value is u201CFALSEu201D. A file is created for each user of the R/3 System group in the u201Csapnamesu201D subdirectory of the transport directory. Except some of the operating system,the name of the user is the name of the file. It is very important to remember hat the special characters or length of the file name could cause problems. If all the R/3 Systems in the transport group have at least Release level 3.0.; TP program is efficient to handle this problem. The user names are modified to create file names that are valid in all operating systems and the real user names are stored in a corresponding file.
    Though we have seen so many parameters, for the minimum configuration the following two parameters are very important.
    TRANSDIR: specifies the name of the common transport directory. The following is a typical example from TPPARAM of Unix as we have seen before.
    transdir = /usr/sap/trans/
    DBHOST: contains the name of the database host. In Windows NT environment, this is the TCP/IP host name. The following is an example in Unix:
    DEV/dbname = DEV
    DEV/dbhost = sap9f
    DEV/r3transpath = /usr/sap/DEV/SYS/exe/run/R3trans
    For TP, to control u2018Start and Stopu2019 command files and database in R/3 the following important parameters are specified in TPPARAM:
    Parameters for the tp Function u201CPUTu201D: LOCK_EU (boolean) default value is u201CTRUEu201D. Though from version 3.1 onward the tp put command is used seldom in cts process still it is important to know how this parameter works. When u201Ctp putu201D is used, it changes the system change option . If the parameter is set to u201CFALSEu201D nothing gets changed. If the parameter is set to u201CTRUEu201D, the system change option is set to u201CObjects cannot be changedu201D at the beginning of the call, and gets changed back to its previous value at the end of the call. The u201Ctp putu201D command will give the exact status of the locking mechanism.
    LOCKUSER (used as boolean value): Default value is u201CTRUEu201D. This parameter is about the user login while tp put call is executed. If this parameter is set to u201CFALSEu201D, no locking mechanism for the users takes affect. If this parameter is defined as u201CTRUEu201D then a character is set in the database level; so only DDIC and SAP* can log on to the system. Users that have already logged on are not affected (this is a reason for activating the parameters STARTSAP and STOPSAP). The charactertor is removed at the end of the call, and all the users can log on to the SAP R/3 System again.
    STARTSAP: Default value is u201D u201C.or u201CPROMPTu201D for Windows NT . This parameter is used by TP to start an R/3 System. It is not necessary for the clients to make tp start and stop R/3 system..
    STOPSAP: Default value is u201D u201Cor u201CPROMPTu201D for Windows NT. TP uses this parameter to stop an R/3 System.
    STARTDB: Default value is u201D u201C. TP uses the value of this parameter to start the database of an R/3 System.
    The parameter is not active under Windows NT.
    STOPDB: Default value is u201D u201C. TP uses the value of this parameter to stop the database of an R/3 System.
    This parameter is not active under Windows NT.
    The above parameters in UNIX can be used as following:
    STARTSAP = startsap R3
    STOPSAP = stopsap R3
    STARTDB = startsap db
    STOPDB = stopsap db
    In Windows NT:
    STARTSAP =
    $(SAPGLOBALHOST)\sapmnt\$(system)\sys\exe\run\startsap.exe
    R3 <SID> <HOST NAME> <START PROFILE>
    STOPSAP =
    $(SAPGLOBALHOST)\sapmnt\$(system)\sys\exe\run\stopsap.exe
    R3 <SID> <HOST NAME> <INSTANCE> <PROFILE PATH + Instance profile>
    The parameters STARTDB and STOPDB are not active under Windows NT.
    Parameters for the tp function u201CCLEAROLDu201D
    DATALIFETIME (Numeric): Default value is u201C200u2033. When the data file has reached a minimum age, it is moved to the subdirectory old data with tp check. tp clearold all. The life span of the data files in the data sub directory can be set in days with this all, parameter.
    OLDDATALIFETIME (Numeric): Default value is u201C365u2033. When a file located in the olddata subdirectory is no longer needed for further actions of the transport system and has reached a minimum age, it is removed with tp check.all, tp clearold all. The minimum age in days can be set with this parameter.
    COFILELIFETIME (Numeric): Default value is u201C365u2033. This parameter is used just like DATALIFETIME parameter.
    LOGLIFETIME (Numeric): Default value is u201C200u2033. This parameter applies to the life span of the log files. When the log files in log subdirectory is no longer needed for the transport system and has reached a minimum age, it is deleted with the calls tp check.all, tp clearold all. The minimum age in days can be defined with this parameter.
    The Three Key Utilities of the CTS system (TP, R3trans and R3chop):
    TP: Earlier in this chapter we have seen the objectives of TP. The TP transport control program is a utility program that helps the user to transport objects from one system to another. TP program is the front-end for the utility R3trans. TP stands for u201CTransports and Putsu201D. To make the TP work successfully the CTS system needs to be correctly configured. The following steps are very important for TP to run properly.
    The transport directory /usr/sap/trans must be installed and NFS mounted to all the systems in the CTS pipe line.
    RDDIMPDP program must be running (event driven is recommended) in each client. RDDIMPDP can be scheduled in the background by executing RDDNEWPP or RDDPUTPP. Use the tp checkimpdp <sap sid> command in /usr/sap/trans/bin directory as <sid>adm user to check RDDIMPDP program.
    Use the tp connect <sap sid> command in /usr/sap/trans/bin directory to see whether the tp program is connecting to the database successfully or not. To run TP command the user has to logon as <sid>adm in source or target system.
    The R/3 Systems in the CTS pipeline must have different names.
    The Global CTS Parameter File TPPARAM must be correctly configured.
    The source system (for the export) and target system ( for the import) must have at least two background work processes. TP always schedules the C class job, so if all the background jobs are defined as A class job then there will be problems in transport steps.
    Important Tips :.It is always better to have the up to date TP version installed in your system. A user can ftp a current version of TP from SAPSERV4 of SAP. Though R3trans and other utility programs can be used to do the transport, it is recommended to use TP whenever possible for the following reasons..
    The exports and imports are done separately using TP program. For example: when a transport is released from the system, the objects are exported from the source database to the operating system and then the import phase starts to transport those objects to the target system.
    TP takes care of the order of the objects. The order, that was followed to export the objects; the same order will be followed to import them to the target database.
    The TP command processes all change requests or transports in the SAP system buffer that have not yet been imported successfully. All the import steps are executed automatically after TP calls R3trans program to execute the following necessary steps:
    Dictionary Import: ABAP/4 dictionary objects will be imported in this step.
    Dictionary Activation: Name tabs or runtime descriptions will be written inactively. The R/3 system keeps running until the activation phase is complete. The enqueue modules are the exceptions in the running phase. After the activation of new dictionary structure the new actions are decided to get the runtime objects to the target system.
    Structure conversion: If necessary the table structure is changed in this phase.
    Move Nametabs: The new ABAP/4 Dictionary runtime objects which were inactive up to now are moved into the active runtime environment in this process. The database structures are adjusted accordingly. From the first step to the Main import step inconsistencies can occur to the R/3 system. After the main import phase all the inconsistency ca be solved.
    Main import with R3trans: All the data are imported completely and the system comes to a consistent state.
    Activation of enqueue-objects: The enqueue-objects cannot be activated in the same way as the objects of the ABAP/4 Dictionary, so they have to be activated after the main import in this step. They are then used directly in the running system.
    Structure Conversion of match codes, Import application defined objects, versioning and execution of user defined activities are some of the steps after activation of enqueue-objects. The next step is generation of ABAP/4 programs and screens, where all the programs and screens associated with the change request are generated. When all the import steps are completed successfully, the transport request is removed from the import buffer.
    It is recommended by SAP to schedule regular periods for imports into the target system (e.g. daily, weekly or monthly). Shorter periods between imports are not advisable. The transport to production should not be done in the off hours when the users are not working
    TP can be started with different parameters. The u201Ctp helpu201D command can help user to generate a short description about the use of the command.
    The following are the some important commands of TP:
    For export:
    tp export <change request>: The complete objects in the request from the source system will be transported. This command also used by SAP System when it releases a request.
    tp r3e <change request>: R3trans export of one transport request.
    tp sde <change request>: Application defined objects in one transport request can be exported.
    tp tst <change request> <SAP system >: The test import for transport request can be done using this command.
    tp createinfo <change request>: This command creates a information file that is automatically done during the export.
    tp verse <request>: This command creates version creates versions of the objects in the specified request.
    To Check the transport buffer, global parameter file and change requests:
    tp showbuffer <sid>: Shows all the change requests ready to be imported to the target system.
    tp count <sid>: Using this command users can find out the number of requests in the buffer waiting for import.
    tp go <sid>: This command shows the environment variables needed for the connection to the database of the <sid> or target system.
    tp showparams <sid>: All the values of modifiable tp parameters in the global parameter file. The default value is shown for parameters that have not been set explicitly.
    To import the change requests or transports:
    tp addtobuffer <request>.<sid>: If a change request is not in the buffer then this command is used to add it to the buffer, before the import step starts.
    tp import all <sid>: This command imports all the change requests from the buffer to the target system.
    tp put <sid>: The objective of this command is same as u201Ctp import all <sid>u201D, but this command locks the system. This command also starts and stops the SAP system, if the parameters startsap and stopsap parameters are not set to u201D u201C.
    tp import <change request> <sid>: To import a single request from the source system to target system.
    tp r3h <change request>| all <sid>: Using this command user can import the dictionary structures of one transport or all the transport from the buffer.
    tp act <change request>|all <sid>: This command activates all the dictionary objects in the change request.
    tp r3i <change request> | all <sid>: This command imports everything but dictionary structures of one.
    tp sdi <change request>|all <sid>: Import application-defined objects.
    tp gen <change request>|all <sid>: Screen and reports are generated using this command.
    tp mvntabs <sid>: All inactive nametabs will be activated with this command.
    tp mea <change request>|all <sid>: This command will activate the enqueue modules in the change request.
    When you call this command, note the resulting changes to the import sequence.
    Additional tp utility options:
    tp check <sid>|all (data|cofiles|log|sapnames|verbose): User uses this command to find all the files in the transport directory that are not waiting for imports and they have exceeded the minimum time specified using the COFILELIFETIME, LOGFILELIFETIME, OLDDATALIFETIME and DATALIFETIME parameters of TPPARAM file.
    tp delfrombuffer <request>.<sid>: This command removes a single change request from the buffer. In case of TMS, the request will be deleted from the import queue.
    tp setstopmark <sid>: A flag is set to the list of requests ready for import into the target system. When the user uses the command tp import all <sapsid> and tp put <sapsid>, the requests in front of this mark are only processed. After all the requests in front of the mark have been imported successfully, the mark is deleted.
    tp delstopmark <sid>: This command deletes the stop mark from the buffer if it exists.
    tp cleanbuffer <sapsid>: Removes all the change requests from the buffer that are ready for the import into the target system.
    tp locksys <sid>: This command locks the system for all the users except SAP* and DDIC. The users that have already logged on are not affected by the call.
    tp unlocksys <sid>: This command unlocks the system for all the users.
    tp lock_eu <sid>: This command sets the system change option to u201Csystem can not be changedu201D tmporarily.
    tp unlock_eu <sid>: This command unlocks the system for all the changes.
    tp backupall <sid>: This command starts a complete backup using R3trans command. It uses /usr/sap/trans/backup directory for the backup.
    tp backup delta <sid>: Uses R3trans for a delta backup into /usr/sap/trans/backup directory.
    tp sapstart <sid>: To start the R/3 system.
    tp stopsap <sid>: To stop the R/3 system.
    tp dbstart <sid>: To start the database.
    tp dbstop <sid>: To stop the database.
    Unconditional modes for TP: Unconditional modes are used with the TP program and these modes are intended for the special actions needed in the transport steps. Using unconditional mode user can manipulate the rules defined by the workbench organizer. The unconditional mode should be used when needed, otherwise it might create problems for the R/3 system database. Unconditional mode is used after the letter u201CUu201D in the TP command. Unconditional mode can be a digit between 0 to 9 and each has a meaning to it. The following is a example of a import having unconditional mode.
    tp import devk903456 qas client100 U12468
    0: Called a overtaker; change request can be imported from buffer without deleting it and then uncoditional mode 1 is used to allow another import in the correct location.
    1: If U1 is used with the export then it ignores the correct status of the command file; and if it is used with import then it lets the user import the same change request again.
    2: When used with tp export, it dictates the program to not to expand the selection with TRDIR brackets. If used in tp import phase, it overwrites the originals.
    3: When used with tp import, it overwrites the system-dependant objects.
    5: During the import to the consolidation system it permits the source systems other than the integration system.
    6: When used in import phase, it helps to overwrite objects in unconfirmed repairs.
    8: During import phase it ignores the limitations caused by the table classification.
    9: During import it ignores that the system is locked for this kind of transport.
    R3trans: TP uses R3trans program to transport data from one system to another in the CTS pipeline. efficient basis administrator can use R3trans directly to export and import data from and into any SAP systems. Using this utility transport between different database and operating system can e done without any problems. Different versions of R3trans are fully compatible with each other and can be used for export and import. The basis administrator has to be careful using R3trans for different release levels of R/3 software; logical inconsistency might occur if the up to date R3trans is not used for the current version of R/3 system.
    The syntax for using the control file is following:
    R3trans [<options>] <control file> (several options used at the same time; at least one option must be there)
    For example: R3trans u2013u 1 u2013w test.log test
    In the above example a unconditional mode is used, a log file u201Ctest.logu201D file is used to get the log result and a control file u201Ctestu201D, where the instructions are given for the R3trans to follow. The user needs to logon as <sid>adm to execute R3trans.
    The following options are available for the R3trans program:
    R3trans -d : This command is used to check the database connection .
    R3trans -u <int>: Unconditional mode can be used as we have seen in the above example.
    R3trans -v : This is used for verbose mode. It writes additional details to the log file
    R3trans -i <file>: This command directly imports data from data file without a control file.
    R3trans -l <file>: This provides output of a table of contents to the log file.
    R3trans -n : This option provides a brief information about new features of R3trans.
    R3trans u2013t: This option is used for the test mode. All modifications in the database are rolled back.
    R3trans -c <f1> [<f2>]: This command is used for conversion. The <f1> file will be copied to <f2> file after executing a character set conversion to the local character set.
    Important tips: Do not confuse the backup taken using R3trans with database backup. The backups taken using R3trans are logical backups of objects. In case something happens to the SAP system these backups can not be used for recovery. R3trans backups can be only used to restore a copy of a particular object that has been damaged or lost by the user.
    R3trans -w <file>: As we have seen in the above example this option can be used to write to a log file. If no file is mentioned then trans.log is default directory for the log.
    R3trans also can be used for the database backup.
    R3trans u2013ba: This command is used for a complete backup. we will see in the next paragraph how to use
    the control file for the backup.
    R3trans u2013bd: This command is used for a delta backup if the user does not want a complete backup.
    R3trans u2013bi: This option will display backup information.
    The following are some of the examples of control files:
    We have already learned how to use a command for the logical backup of the objects in the database. To get a complete backup the following example control file can be used.
    backup all
    file = /usr/sap/trans/backall
    The option u201Cfile = u2026u201D is the name of the directory into which the data files are to be written. If you are taking a complete backup of DEV system then the backup file is going to look like u201CDEV.A000.bcku201D the next complete
    Reward if useful
    Regards
    divya

  • How to do transports in BI7..

    WE HAVE A REQUIREMEN to transport a new report from dev to prod. the data flow is we have an infoset defined on 1 cube and 4 dso's , out of these 5 infoproviders 1 dso is not there in production, the info set is not there in production and the report is also not there.
    Can anyone help how to transport that report from dev to prod , what should be the order in which we have to transport.
    how to tranport the infoset and dso and the report.
    also how is the data source transported..
    we have all the objects in production except the report , infoset and the dso.
    Please help me how to do the transports , step by step with all the tcodes involved

    1. Create your request in SE01 in R/3 (R/3 transports) & in BI (for BI Transports).
    Assign the request while saving the object. Package will be assigned to you by Basis.
    Or,
    Go to Object Directory Entry (search in menu bar options) & change this from $TMP to the package that you have been assigned by Basis. Then create the transport request from as well.
    2. Customizing request: this is a Client (SAP Client) specific request
    - Normally in Dev we may have 2 clients for the same system.
    - One client is used for Customizing the changes & the other may contain some test data
    - The customizing request is copied between these 2 clients
    This is then transported to the next system
    Workbench request: This is not client specific & is also used for transports
    3. Normally package is assigned by Basis for particular projects. Please ask the basis team for this.
    We can create this, but normally basis creates this.
    4. Datasource in R/3 Dev needs to be transported to R/3 Prod. What is the custom data source that you have made?
    You may have to see if the user postings have to be stopped for this or not.
    After you have activated the data source in R/3 Dev, replicate this in BI Dev & capture the active version in a transport in BI Dev as well.
    After transporting the data source to R/3 Prod, replicate this in BI.
    Send the transport of the active version from BI Dev to BI Prod after replication.

  • Transport Best Practices - Cumulative Transports

    Hi All,
        I am looking for a some sort of authoritative guide from SAP on ECC Transport Best Practices, especially around merging/combining/accumulating multiple transports into fewer ones. We are a very large project, but we haven't figured out the CVSs like ChaRm so we still deal with individual transports.
    The reason I am asking this is that we some development leads on our project that insist that ALL transports that leave Development system must be imported into Production.  They claim this the SAP best practice.  An SAP consulting review of our system also left a vague note that "We are orphaning transports". This could mean that 1. either we are not importing all the stuff that leaves Dev system 2. or we are not keeping track of our code changes across all environments. Proponents of "All transports must get to PRD" are interpreting this as "1". 
    I have my team cumulate transport for subsequent changes into newer transport and only take the newest Transport to Production.  The continuous rolling of old Transport into new one using SE01 "Include Objects" options ensures that all changes part of current development are in a single TR. Fewer transports mean fewer housekeeping across all systems, and less chances of something going out of sequence or missed out. This is for Workbench Transports only. I understand Config transports could get a little tricky with rolling in.
    If you can't point me to a link, what is your take on "Send everything to Prod" vs. "Combine changes into fewer Transports"?  I can't think of any software packaging methodology that suggests putting everything, including some junk/crap, into production build. I have looked at SAP enhancement packs, SPS, and Notes. I haven't found any evidence of  SAP including older buggy code in what it releases to its customers.
    Thank you all!

    Jānis, Christian,
        I think we are all on the same page.  Let me clarify my specific scenario a little bit more.
    We are about 15 ABAP developers team, for production support. We don't do huge changes. Our code updates are mostly limited to a handful of objects (average 2 - 5).  We often have multiple iterations to same objects. For large scale development objects, this approach may not work very well. Those should really utilize SolMan or other CVS tools.
    Here is How I have my team putting together final transport.
    step 1. Change is done to object X. Transport #1 created and released after standard checks.
    step 2. More change is needed for object X.  Transport #2 Started.  Transport #1 is brought in using SE01 include objects at the main task. Changed objects are in lower tasks. This way, I can tell what was brought over, and what really changed this time.  Releases of the Transport #2 inspects all objects and insures all objects from #1 are included in #2, and that there are no other changes between #1 and #2 to any of the objects.  This is very easy check mostly from Version History.
    Step 3. More changes needed to object X and Y. Transport #3 started.  Transport #2 brought in at main task. Same check from Step 2 is done to ensure no other changes exist.
    step 4....6. Step 6 ended at Transport #6.  All changes verified in QA system.
    Only the Transport #6 needs to be sent to Production since previous Transports 1 to 5 were rolled into #6.
    Jānis, the deletions will be covered automatically just like standard SAP.  No special manual steps needed.
    Christian,
       The transport of copies works in similar way.  Only thing different is that the Main/cumulative transport is released at very last, possibly after QA tests have been confirmed.
    I had thought about doing Copies vs. cumulative.  I preferred to go with having Transport in QA already at the time of approval. I would have hard time explaining to our client why a Transport was released out of Dev After all tests were done and signed off.
    However, the Copies have advantage that intermediate versions or parallel changes of same objects are easily recognized upfront vs. us having to check each transport before release.
    Jānis,  the "copies of Transport" also creates extra versions in version history just like manually adding objects in a transport. 
    My quick analysis between copies and cumulative only came with one different that 'copies' have a different flag/attribute in Table E070.  I am sure the flag means something to SAP, but for versioning history, it makes no differences.
    Regardless of if I cumulate or not, based on your experiences, have you come across notion that everything that leaves Dev must get to PRD as SAP Best Practice? What would your reply  be if someone insists that is the only way to move code to production?

  • Detailed step-by-step technicals - INFOPACK & PC

    Hi SDNers,
    *POINTS GURANTEED
    I am working with process chains in my proj. Theres an ODS - ZODSA, for which no load is scheduled for the past few months. Now it is needed to load this infopack & that too creating a PC for this & attaching it to the main PC.
    Could you pls give me a detailed step-by-step technicals in carrying out the following?
    1) Migrate existing infopackage for ZODSA from dev to quality
    2) (LEG) create new subchain to execute, then activate this load - start condition should be immediate, no server name
    3) Attach this new chain to Z_TRANS_ALL, after the master data loads, in parallel with ZSALES_PC in dev and test
    4) Migrate and test PC changes,
    5) Migrate IP and chain to Production and schedule chain for daily full load.
    POINTS GURANTEED

    1) Migrate existing infopackage from dev to quality.
    RSA1 - Transport Connection - Grouping - Only Necessary - Collection - Manual
    Infopackages - Search for the Infopackage - Transfer Selection - Package - Select - Assign Transportable Package - Assign Request.
    2) In dev, a new subchain is to be created to execute, then activate this load - start condition should be immediate
    How to keep the start variant as changable in Qual?
    Usually, start variants will be open in Qual for changing scheduling timings or others.
    3) In dev, Attach this new chain to the main chain
    If you know where to add the new chain to the main chain - RSPC - Edit Mode - Processing Types - Local Chain - Read the new chain - Connect to pre and succesor - Activate
    In the start variant of this local chain - Select " Start Using API........."
    4) Then to transport BOTH THE CHAINS from Dev to Qual and test there also
    Click on the truck icon from RSPC itself and foll above steps.
    You can just select process types and infoapackages if all others have already been transported.
    Pls give me a detailed transport steps.
    5) Finally the infopackage + BOTH THE CHAINS are to be moved to prodn, right?
    Yes

  • Web Interfaces can be displayed incorrectly in QA system after transported

    Dears,
      I transported Web Interfaces for BPS from D system to Q system. I activate the BSP application in transation SE80 and activate the service node in transaction SICF in Q system. However, the BSP can't be displayed correctly with NO_METADATA_AVAILABE error.Also, I test the HTTP service in Q system, it works well. The transport steps I followed:
    1) change the package of BSP application (SE80 D system)
    2) write transport entry to new request (SE80 D system)
    3) switch the correspoding service node to new object directory entry (SICF D system)
    4) transport the request to Q system
    5) activate BSP application (SE80 Q system)
    6) activate BSP service node (SICF Q system)
    Am I missing configuring something? or completely wrong steps? Any suggestions are appreciated.
    Best regards,
    Gerald

    I resolved it myself

Maybe you are looking for