Export and Import of mappings/process flows etc

Hi,
We have a single repository with multiple projects for DEV/UAT and PROD of the same logical project. This is a nightmare for controlling releases to PROD and in fact we have a corrupt repository as a result I suspect. I plan to split the repository into 3 separate databases so that we have a design repos for DEV/UAT and PROD. Controlling code migrations between these I plan to use the metadata export and subsequent import into UAT and then PROD once tested. I have used this successfully before on a project but am worried about inherent bugs with metadata export/imports (been bitten before with Oracle Portal). So can anyone advise what pitfalls there may be with this approach, and in particular if anyone has experienced loss of metadata between export and import. We have a complex warehouse with hundreds of mappings, process flows, sqlldr flatfile loads etc. I have experienced process flow imports that seem to lose their links to the mappings they encapsulate.
Thanks for any comments,
Brandon

This should do the trick for you as it looks for "PARALLEL" therefore it only removes the APPEND PARALLEL Hint and leaves other Hints as is....
#set current location
set path "C:/TMP"
# Project parameters
set root "/MY_PROJECT"
set one_Module "MY_MODULE"
set object "MAPPINGS"
set path "C:/TMP
# OMBPLUS and tcl related parameters
set action "remove_parallel"
set datetime [clock format [clock seconds] -format %Y%m%d_%H%M%S]
set timestamp [clock format [clock seconds] -format %Y%m%d-%H:%M:%S]
set ext ".log"
set sep "_"
set ombplus "OMBplus"
set omblogname $path/$one_Module$sep$object$sep$datetime$sep$ombplus$ext
set OMBLOG $omblogname
set logname $path/$one_Module$sep$object$sep$datetime$ext
set log_file [open $logname w]
set word "PARALLEL"
set i 0
#Connect to OWB Repository
OMBCONNECT .... your connect tring
#Ignores errors that occur in any command that is part of a script and moves to the next command in the script.
set OMBCONTINUE_ON_ERROR ON
OMBCC "'$root/$one_Module'";      
#Searching Mappings for Parallel in source View operators
puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Searching for Loading/Extraction Operators set at Parallel";
puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Searching for Loading/Extraction Operators set at Parallel";
puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
foreach mapName [OMBLIST MAPPINGS] {
foreach opName [OMBRETRIEVE MAPPING '$mapName' GET TABLE OPERATORS] {
foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT)] {
if { [ regexp $word $prop1] == 1 } {
incr i
puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
OMBCOMMIT;
foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
if {[regexp $word $prop2] == 1 } {
     incr i
puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
OMBCOMMIT;
foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET DIMENSION OPERATORS ] {
foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
if {[regexp $word $prop1] == 1 } {
     incr i
puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
OMBCOMMIT;
foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
if {[regexp $word $prop2] == 1 } {
     incr i
puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
OMBCOMMIT;
foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET CUBE OPERATORS ] {
foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
if {[regexp $word $prop1] == 1 } {
     incr i
puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
OMBCOMMIT;
foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
if {[regexp $word $prop2] == 1 } {
     incr i
puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
OMBCOMMIT;
foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET VIEW OPERATORS ] {
foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
if {[regexp $word $prop1] == 1 } {
     incr i
puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
OMBCOMMIT;
foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
if {[regexp $word $prop2] == 1 } {
     incr i
puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
OMBCOMMIT;
if { $i == 0 } {
          puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
          puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Not found any Loading/Extraction Operators set at Parallel";
          puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
          puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Not found any Loading/Extraction Operators set at Parallel";
     } else {
          puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
          puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Fixed $i Loading/Extraction Operators set at Parallel";
          puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
          puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Fixed $i Loading/Extraction Operators set at Parallel";
close $log_file;
Enjoy!
Michel

Similar Messages

  • Separate Distribution Monitor Export and Import Processes on Multiple Machines

    Hi,
    Would you kindly let me know whether it is possible (means officially supported way) to run Distribution Monitor Export and Import processes on different machines?
    As per SAP note 0001595840 "Using DISTMON and MIGMON on source and target systems", it says as below.
    > 1. DISTMON expects the export and import to be carried out on the same server
    I think it means that export and import processes for the same tables must be run on the same machine, is it correct? If yes, Export on the machine A, and then Import those exported data on the other machine B is not the officially supported way... (However, I know it is technically possible.)
    Kind regards,
    Yutaka

    Hi Yutaka,
    Point no. 2 & 3 clarify the confusion. However let me explain it briefly:
    Distribution Monitor is used basically in case of migration of large SAP systems (database). It provides the feature to increase parallelism of export and import, distributing the process across available systems.
    You have to prepare the system for using DistMon. A common directory needs to be created as"commDir" and in case you use multiple systems for executing more number of processes of export and import then that "commDir" should be shared across all those systems.  And this is what the Point no.1 in KBA 1595840 mentions about. Distribution Monitor will run both the export and import process from the machine which is prepared for using DistMon and DistMon itself will control the other processes i.e. MigMon. No need to start separate MigMon.
    For example: You are performing a migration of SAP system based on OS:AIX and DB:DB2 to  OS: HP-UX and DB: Oracle. You need to perform the export using DistMon and you are having 4 Windows servers which can be used for parallel export/import. Once you have prepared the system for DistMon which hosts the "commDir" you'll have to provide the information of involved host machines in the "distribution_monitor_cmd.properties" file. Now when DistMon is executed it will distribute the export and import process across the systems which were defined in "distribution_monitor_cmd.properties" file automatically.
    Best regards,
    SUJIT

  • Is it possible to replace masters with versions without going through the process of exporting and importing?

    Is it possible to overwrite the master files with the versions without having to go through the entire process of exporting and importing back the images?
    Thnx in advance

    A small refinement if you shoot RAW. Remember, no RAW file can ever be edited or rewritten. Thus, in the old days, the flow was:
    RAW>TIFF/PSD and you ended up with two large files. (This is why some folks used to delete the RAW Masters.)
    With Aperture the flow is something like:
    RAW+Version(s) Only the Version is only 30 KB. (There is, of course, additional overhead of Thumbs and Previews, but, depending on your settings, this is only a fraction of the size of a TIFF or PSD.)
    You question goes right to the heart of the concept of a non-destructive workflow. With Aperture you always preserve your original Master file; the only things that are ever at risk are your edits (called Adjustments). Ironically, this is even more valuable if you shoot JPEG or TIFF as those files could be rewritten.)
    Aperture goes to great lengths to give you the best of both worlds; you always have the Master to revert to and your "edits" take up very little disk space. And you don't have to waste time trying to sync folders of RAW, TIFF, and JPEG. Similarly, you never have to worry about resizing images, changing color spaces, or any number of other things that get locked into an image file when you use a destructive workflow.
    Remember, Aperture will happily Export files in any format you may need. Export 'em, use 'em, delete 'em - your Masters are still safe in Aperture.
    DiploStrat

  • Export and Import in Data warehouse

    Hi,
    I have built a data warehouse project with few dimensions and a cube on a server machine, say server1. Now i have another server machine on which i want to develop this project for bulk production.
    I want to import all the developments from server1 to new server machine, say server2, To save my time. I exported data from server1 using export functionality and imported the same file to server2 using import function. My questions is.......
    Is it possible to use all the imported objects like cube, dimension, mappings, tables and sequences in the same way in server2 as i was using in server1. If yes then what are the necessary configurations, i have to made.???? because imported dimensions are not able to deploy in new server2 machine giving different errors especially invalid location and invalid table errors for mapping and dimensions.
    thanks in advance
    imran

    you can sue the MDL file on your new server but you have to do the following:
    1. Register locations.
    2. Configure the mapping to use the new locations registered.
    4. Configure Control Center locations.
    Register the location in Control center
    Deploy the mappings/Process flows
    The Invalid location error is occuring becaue the location is not configured and registered in your connection explorer.

  • Export and Import of Application problem

    Hi,
    I had my application developed in Windows environment. First time i had full application export and import on Solaris box. The application was successfully installed on Solaris box and i could login into application. The login process was working but all of my queries/reports were not displaying any values. All i can see is number of rows retrieved for example 1-20 etc. I repeated the export/import 3-4 times but results were same.
    On Solaris box i opened one of the reports and started playing with it. Luckily on "Reports Attribute" tab i unchecked/checked the show column checkbox. The report started working. Likewise i did for other reports. Just to verify this i repeated the Export/import of application and unchecked/checked the show column and reports started working. I repeated it at least 3 times.
    Our application is going to expand everyday and on every import we cannot repeat this step for every report.
    Please let me know what steps are missing in Import or are they any manual Post-Import steps ?
    Regards
    Hamayun

    Hi Scott,
    Today i applied the patch and now its 1.6.1.02. After applying the patch i imported application once again.
    I was able to login into application. When i executed a report the system navigated to required tab but did not show any data. Then i performed following steps:
    1.Selected Edit page(xx).
    2.Selected region in the selected page.
    3.Select Report Attributes tab from region.
    4.In Report Column Attributes section the "Show" option was Checked for all the columns which i want to display.
    5. i unchecked only ONE COLUMN and CHECKED again the same column. This was done just to apply changes to report.
    6. Clicked on Apply Changes to save the changes.
    7. Executed the application and the report started working.
    8. Likewise i repeated step 5 for all reports in the application.
    I also tried another option.
    -navigated to Reports Attribute tab and made a change just to save the changes on tab. After saving the page the report started working.
    Your help is greatly appreciated.
    Regards,
    Hamayun

  • Query on Database Export and Import

    Hi Techies,
    Currently we are running our Database on Oracle 10g and SAP on4. and OS is HP UX 11.11.
    We have a plan to migrate our HW from PA-RISC to Itanium and at the time of Production migration we are planning to use Export and Import method to get free space.
    Our plan is as below:
    We will not touch the original Production, Just restore the DB into new server.
    And post restore we will create the space on new server equallent to our DB size.
    Then will perform DB export from the new system to the null space
    Then Import the DB into same system.
    Here my queries are:
    1) Is it possible to export and Import the Database from/to null space?
    2) We have 2T size of DB, And good resources like 32G Ram, 12 CPUs etc. How much time can be expected to perform Export and Import?
    3) What are the challenges we can expect?
    4) Minimum how much of free space we can expect with this option?
    Regards,
    Nick Loy

    So with test runs I can expect rapid speed in DB export and Import (1T/H)........If I have good system then Database export and Import gets complete within 2 hrs (Database size is 2T).
    Well 1tb is at the very top of expectations here, you should be careful. I did an export/import of a 1.5tb database of an ERP system lately. We did parallel export (40 processes) / import (20 processes) using distmon, source was HP IA64, target Linux x64_64. The disks were midrange SAN storage systems on both sides.  After tuning we managed to do in 6-7hrs.
    But in your case, if you only have one system, this could mean you have to drop the source db first and then recreate the target db on the same disks. The creation of the 1-2tb database files alone can take up more than 1 hour , besides that you don't have an easy fallback.
    If you have a test system that is comparable from size and hardware perspective, then i suggest you try a test export to get a feeling for it.
    What about the Online re-org of database? What would be the best way to get free space within minimum downtime?
    Theoretically you should be able to gain more or less the same amount of space doing online reorgs. The advantage is less downtime, the downside is the reorgs will be running over a longer time period and put additional load on the system.
    Cheers Michael

  • Automate exports and imports (file names)

    Hi all,
    I will like to do a simple export and import into another schema in another database.
    I am very comfortable with the process but all i require is how to set the dump file and log file names to be the date and name of the database being exported and imported.
    For example when I do the export I want the dump file and logfile names to be exp_ORCL_230805.dmp and exp_ORCL_230805.log for database ORCL exported on 230805.
    For the import I want the logfile name to be imp_TEST_230905.log.
    I am doing this export and import on a windows xp client but the database reside on UNIX AIX boxes.
    Thanks alot

    Set up ORACLE_SID before you start the imp/exp session on the client, then you can use file=exp_%ORACLE_SID%_%DATE%.dmp etc. (Note that if your windows locale has a date format w/ whitespace you might have to write a script to get rid of it..)

  • How to export and import the Portal Components

    Hi,
    I had developed few components like (Forms, Reports, Frames etc.,) in Oracle Portal release 2. Now, I need to make available all the components on my Client's machine. I like to know how to export and import these components.
    If any can help in this, it will be a great help for me...
    Thanks in Advance...

    The migration of any Portal DB provider components, has the following steps.
    Assume that you are migrating a form component form_X.
    1) Click the export link of the form_x.
    (this will take you to create transport screen).
    2) Specify a name of the transport set which you are going to create. (eg., "My form_X transportset")
    (Since the objects are migrated using a container which is in other words called as 'Transport Set').
    3) In that select the export now button.
    (this will mark your form component form_X for export).
    4) Verify that the status of the transport is 'Extract Complete'
    (check with the browse transport screen in Export/Import Portlet under Admin tab).
    5) Now, the object information is extracted into transport set tables used by Portal Export/Import Utility.
    6) Go to that particular transport set, (browse transport) and download the script available.
    7) That script will lead you to various steps,
    (create a dump ).
    8) In the target instance, import that dump using the same script again.
    9) Merge the contents of the dump (form_X information) into any of the application.
    (Import option => in the export/import portlet under admin tab).
    10) Verify that the status of the transport set is 'Merge Complete'.
    Now, the object is migrated and ready for use.
    (Here are explaining the export/import process in detail.
    You can refer the export/import manuals.
    http://portalstudio.oracle.com/pls/ops/docs/FOLDER/COMMUNITY/INTERNALPRODDEVFOLDER/TECHREADINESS/CM_PUBLISHING/EXPIMP/EXPORT_IMPORT_LOB0.PPT

  • URGENT...Help on User Level export and import

    User level export and import -
    In my db there are more than 20 users, i would like export only 2 users and they have constraints to other user tables in db.
    what are the precautions I need to take before exporting ???
    1. Disabling all constraints etc, .
    2. Synonymns, Procedures , Triggers etc.
    3. what are the parameters setting I should use for the export.
    For Importing back into DB (same users) - What are the sequences/steps to follow..
    1. what are the parameters setting I should use for the Import.
    2. Precautions like Disabling Constraints...
    3. Special actions for Synonyms/ Procedures.
    Using Oracle 8.1.7 on Solaris.. Total DB Size is 13.3 GB
    Any help will be appreciated....

    See to it that u have the DBA PRIVS to exp/imp.
    1.Grant export full database /import full database privileges to the user doing this.
    2.If this is the first time u are doing this u have to run the scripts CATEXP.SQL and CATPROC.SQL to create the necessary views,datadictionaries as user SYS/INTERNAL.These instructions are also mentioned in the scripts comments at the beginning .
    3.Also create the necessary tablespaces where u need to direct u'r imp.
    4.select the interactive option where u can interact with
    imp/exp process and u can do this selectively for users/tables/partitions and other preferences.
    5.look out for the version compatibility of your .DMP files

  • Use of export and import in oo abap

    Hello experts,
    We are new to oo abap and using process controlled workflow for BRF.
    There are 3 methods which we are using. /SAPSRM/IF_EX_WF_RESP_RESOLVERGET_AREA_TO_ITEM_MAP, /SAPSRM/IF_WF_AREAGET_RESPONSIBLE_APPROVERS and /SAPSRM/IF_EX_WF_RESP_RESOLVER~GET_APPROVERS_BY_AREA_GUID.
    In a situation we want to use document_id which is created in item_map into get_responsible_approvers method.
    Therefore, we used export and import in these two methods like the following.
    zgc_mock_pdo_guid =  is_document-document_guid. "uncomment later
        export zgc_mock_pdo_guid from zgc_mock_pdo_guid to memory id 'Z_MOCK_PDO_GUID' . "uncomment later
    and
    import zgc_mock_pdo_guid  = zgc_mock_pdo_guid  from  memory id 'Z_MOCK_PDO_GUID' . "uncomment this later
                   CALL METHOD /SAPSRM/CL_WF_APV_FACADE=>RETRIEVE_PROCESS_HISTORY
                     EXPORTING
                       IV_DOCUMENT_GUID = zgc_mock_pdo_guid "ins_guid
                       IV_AGENT_ID      = lv_agent
                       IV_LANGUAGE      = sy-langu
                     IMPORTING
                       ES_PROCESS       = ls_process
    Now we are getting the results as per our requirements. But I want to know whether or not it is supposed to be used in oo abap context or not otherwise I will find another method to do so. Please advise/suggest about the same.
    Thank you.

    Hello Keshav, Suhas.
    Thank you vary much for your reply. I was definitely a great help to understand the concept.
    This ZGC_MOCK_PDO_GUID attribute was already made private. Instead of defining under public section I defined it under attributes tab.
    Thank you for your help though.
    Right now I am trying to explore another way to do so.
    I am trying to get factory instance for tack container using /SAPSRM/IF_CLL_TASKCON_FACTORY interface with method GET_INSTANCE from class /SAPSRM/CL_CH_WD_TASKCONT_FACT.
    Later I am trying to get task container from this interface. Then after I am interested to get SC GUID using GET_BO_GUID method.
    I have been trying to achieve this in following way. But somehow i am getting this GUID empty.
    data :
    temp_task_container_factory type REF TO /SAPSRM/IF_CLL_TASKCON_FACTORY,
    temp_instance type REF TO /SAPSRM/IF_PDO_BO_SC_ADV,
    ins_guid type BBP_GUID, "/SAPSRM/WF_DOCUMENT_GUID,
    ls_process_info TYPE /SAPSRM/S_PDO_WF_PROCESS_INFO,
    temp_handler TYPE REF TO /SAPSRM/IF_PDO_STATIC_META,
    temp_task_container type REF TO /SAPSRM/IF_CLL_TASK_CONTAINER,
    lo_task TYPE REF TO /SAPSRM/IF_CLL_TASKCON_FACTORY.
    CALL METHOD /SAPSRM/CL_CH_WD_TASKCONT_FACT=>GET_INSTANCE
    RECEIVING
    RO_TASK_CONTAINER_FACTORY = temp_task_container_factory "
    CALL METHOD TEMP_TASK_CONTAINER_FACTORY->GET_TASK_CONTAINER
    * EXPORTING
    * IV_TRANS_MODE =
    * IO_MESSAGE_HANDLER =
    * IV_BO_TYPE =
    RECEIVING
    RO_TASK_CONTAINER = temp_task_container
    CALL METHOD TEMP_TASK_CONTAINER->GET_BO_GUID
    RECEIVING
    RV_BO_GUID = ins_guid .
    TRY.
    CALL METHOD /SAPSRM/CL_PDO_FACTORY_SC_ADV=>GET_BUFFERED_INSTANCE
    EXPORTING
    IV_HEADER_GUID = ins_guid
    RECEIVING
    RO_INSTANCE = temp_instance
    ENDTRY.
    CALL METHOD ZCL_WF_APV_FACADE=>RETRIEVE_PROCESS_HISTORY
    EXPORTING
    IV_DOCUMENT_GUID = ins_guid " we are passing ins_guid instead of zGC_MOCK_PDO_GUID this should have area_guid now.
    IV_AGENT_ID = lv_agent
    IV_LANGUAGE = sy-langu
    IMPORTING
    ES_PROCESS = ls_process.
    But the ins_guid ( returned by RV_BO_GUID ) is empty ( The Bold fonted line of code)
    I want to know whether I am doing this instantiation correctly or not or do I have to create class and object here only ?
    Right now I am calling the class, getting get_instance and using its instance temp_task_container_factory i am calling interface /SAPSRM/IF_CLL_TASKCON_FACTORY. Is it correct way? PLease advise. Thank you once again.
    Edited by: abhijitkamatkar on Feb 7, 2012 1:38 PM

  • Client Export and Import

    Hi,
    I am having one doubt regarding Client export and import
    In my organization we are having three system landscape, but the Quality and Production server(In Process)
    Development box is ready to use, and we are having another sand server in different location for our consultant practice
    that system is not in our landscape and that is different location,i want to import that client in to my development server hear
    It is possible to export that data from that Client(scc8) copy that in my development server and Import(scc7) is possible..?
    if possible than what are the pre-requisites to do That Export and Import of client between two different systems,
    what is best procedure to do this Task.......!
    Regards'
    Mohammed.Khan

    >>once export is done successfully than can i copy those exported file in to my file server and then copy to my development >>system trans directory (cofiile,datafile) than can i import those files in my development server with *stms_import* command
    Yes, you have to
    >>than can i go for client import(scc7)
    Yes.
    >>***Configuring the TMS between that sand box and my development server***
    Not required but yes if the sandbox server is not in the landscape you have to create a Virtual System in sandbox and you have to configure the Transport, then only you would be able to create the export file
    Regards
    Subhash

  • Export and Import of Configuration Data

    Hi All,
    I have created a number of custom properties, structures, groups, renderers, layout sets etc and would like to move them from Dev to Test system.
    I have found the following in the help doco:
    "Export and Import of Configuration Data"
    http://help.sap.com/saphelp_nw04/helpdata/en/e1/029c414c79b25fe10000000a1550b0/content.htm
    However I am unable to see its functionality in our portal systems (namely the Actions->Export and Actions->Import functions).
    Our version of Portal is: 6.0.9.6.0
    KM: 6.0.9.3.0 (NW04 SPS09 Patch3)
    Therefore we should be able to see it? no?
    Are there any special settings to see these buttons?
    Cheers,
    Vic

    Thanks for your replies.
    OK, so in otherwords, it is not available in SP09 and only in SP12+.
    So, to "transport" this information I will have to manually reapply thes settings in the Test system!
    Message was edited by: Victor Yeoh

  • Export and Import of APPS and Applsys Schemas in R12

    Hi ,
    Can you please let me know Export and Import of APPS and Applsys Schemas in R12 will cause any issue ?
    Thanks and Regards,
    Jagadeesha.

    Can you please let me know Export and Import of APPS and Applsys Schemas in R12 will cause any issue ?If you mean exporting/importing only those two schema between different instances then this is not supported (due to other schema dependencies) and the only supported way to export those two schema is by exporting/importing the complete database.
    Export/Import Process for Oracle E-Business Suite Release 12 using 10gR2 [ID 454616.1]
    Export/import process for R12 using 11gR1 or 11gR2 [ID 741818.1]
    Thanks,
    Hussein

  • View and Edit properties pages are not woriking after export and import in sharepoint2013

    Hi,
    After performing Export and Import in SharePoint 2013 for a subsite, metadata is missing in document library. 404 error triggers on  clicking both View-properties of an item and /Forms/DisplayDocument.aspx.
    Please help me in this.
    Regards,
    Gurudatta

    While uploading, after browsing the document clicking on OK button giving below error.
    Nothing is generated in Event logs. Below are sharepoint logs.
    07/16/2014 17:32:08.44  w3wp.exe (0x5234)                        0x499C SharePoint Foundation        
     Logging Correlation Data       xmnv Medium   Name=Request (GET:http://siteurl/frmFile4/Forms/CreateDocument.aspx?Mode=Upload&CheckInComment=&ID=963&RootFolder=/QAPortal/PackingMaterial/frmFile4&Source=http://siteurl/frmFile4/Forms/AllItems.aspx?InitialTabId=Ribbon%252EDocument&VisibilityContext=WSSTabPersistence&RootFolder=%252FQAPortal%252FPackingMaterial%252FfrmFile4&IsDlg=1) 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.44  w3wp.exe (0x5234)                        0x499C SharePoint Foundation        
     Request Management             adc7u Medium   Mapping URI from 'http://siteurl/frmFile4/Forms/CreateDocument.aspx?Mode=Upload&CheckInComment=&ID=963&RootFolder=%2FQAPortal%2FPackingMaterial%2FfrmFile4&Source=http%3A%2F%2Fuslsp13%3A1111%2FQAPortal%2FPackingMaterial%2FfrmFile4%2FForms%2FAllItems%2Easpx%3FInitialTabId%3DRibbon%252EDocument%26VisibilityContext%3DWSSTabPersistence%26RootFolder%3D%252FQAPortal%252FPackingMaterial%252FfrmFile4&IsDlg=1'
    to 'http://siteurl/frmFile4/Forms/CreateDocument.aspx?Mode=Upload&CheckInComment=&ID=963&RootFolder=%2FQAPortal%2FPackingMaterial%2FfrmFile4&Source=http%3A%2F%2Fuslsp13%3A1111%2FQAPortal%2FPackingMaterial%2FfrmFile4%2FForms%2FAllItems%2Easpx%3FInitialTabId%3DRibbon%252EDocument%26VisibilityContext%3DWSSTabPersistence%26RootFolder%3D%2... 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.44* w3wp.exe (0x5234)                        0x499C SharePoint Foundation        
     Request Management             adc7u Medium   ...52FQAPortal%252FPackingMaterial%252FfrmFile4&IsDlg=1' 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.44  w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Monitoring                     nasq Medium   Entering monitored scope (Request (GET:http://siteurl/frmFile4/Forms/CreateDocument.aspx?Mode=Upload&CheckInComment=&ID=963&RootFolder=%2FQAPortal%2FPackingMaterial%2FfrmFile4&Source=http%3A%2F%2Fuslsp13%3A1111%2FQAPortal%2FPackingMaterial%2FfrmFile4%2FForms%2FAllItems%2Easpx%3FInitialTabId%3DRibbon%252EDocument%26VisibilityContext%3DWSSTabPersistence%26RootFolder%3D%252FQAPortal%252FPackingMaterial%252FfrmFile4&IsDlg=1)).
    Parent No
    07/16/2014 17:32:08.44  w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Logging Correlation Data       xmnv Medium   Name=Request (GET:http://siteurl/frmFile4/Forms/CreateDocument.aspx?Mode=Upload&CheckInComment=&ID=963&RootFolder=%2FQAPortal%2FPackingMaterial%2FfrmFile4&Source=http%3A%2F%2Fuslsp13%3A1111%2FQAPortal%2FPackingMaterial%2FfrmFile4%2FForms%2FAllItems%2Easpx%3FInitialTabId%3DRibbon%252EDocument%26VisibilityContext%3DWSSTabPersistence%26RootFolder%3D%252FQAPortal%252FPackingMaterial%252FfrmFile4&IsDlg=1) 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.46  w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Authentication Authorization   agb9s Medium   Non-OAuth request. IsAuthenticated=True, UserIdentityName=0#.w|usl\hclspp, ClaimsCount=23 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.50  w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Information Rights Management  5202 Information Information Rights Management (IRM): Requesting user email address is empty. 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.50  w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Files                          ak8dj High     UserAgent not available, file operations may not be optimized.   
    at Microsoft.SharePoint.SPFileStreamManager.CreateCobaltStreamContainer(SPFileStreamStore spfs, ILockBytes ilb, Boolean copyOnFirstWrite, Boolean disposeIlb)     at Microsoft.SharePoint.SPFileStreamManager.SetInputLockBytes(SPFileInfo&
    fileInfo, SqlSession session, PrefetchResult prefetchResult)     at Microsoft.SharePoint.CoordinatedStreamBuffer.SPCoordinatedStreamBufferFactory.CreateFromDocumentRowset(Guid databaseId, SqlSession session, SPFileStreamManager spfstm,
    Object[] metadataRow, SPRowset contentRowset, SPDocumentBindRequest& dbreq, SPDocumentBindResults& dbres)     at Microsoft.SharePoint.SPSqlClient.GetDocumentContentRow(Int32 rowOrd, Object ospFileStmMgr, SPDocumentBindRequest&
    dbreq, SPDocumentBindResults& dbres... 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.50* w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Files                          ak8dj High     ...)     at Microsoft.SharePoint.Library.SPRequestInternalClass.GetFileAndMetaInfo(String
    bstrUrl, Byte bPageView, Byte bPageMode, Byte bGetBuildDependencySet, String bstrCurrentFolderUrl, Int32 iRequestVersion, Byte bMainFileRequest, Boolean& pbCanCustomizePages, Boolean& pbCanPersonalizeWebParts, Boolean& pbCanAddDeleteWebParts, Boolean&
    pbGhostedDocument, Boolean& pbDefaultToPersonal, Boolean& pbIsWebWelcomePage, String& pbstrSiteRoot, Guid& pgSiteId, UInt32& pdwVersion, String& pbstrTimeLastModified, String& pbstrContent, UInt32& pdwPartCount, Object&
    pvarMetaData, Object& pvarMultipleMeetingDoclibRootFolders, String& pbstrRedirectUrl, Boolean& pbObjectIsList, Guid& pgListId, UInt32& pdwItemId, Int64& pllListFlags, Boolean& pbAccessDenied, Guid& pgDocid, Byte& piLevel,
    UInt64& ppermMask, ... 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.50* w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Files                          ak8dj High     ...Object& pvarBuildDependencySet, UInt32& pdwNumBuildDependencies,
    Object& pvarBuildDependencies, String& pbstrFolderUrl, String& pbstrContentTypeOrder, Guid& pgDocScopeId)     at Microsoft.SharePoint.Library.SPRequestInternalClass.GetFileAndMetaInfo(String bstrUrl, Byte bPageView, Byte
    bPageMode, Byte bGetBuildDependencySet, String bstrCurrentFolderUrl, Int32 iRequestVersion, Byte bMainFileRequest, Boolean& pbCanCustomizePages, Boolean& pbCanPersonalizeWebParts, Boolean& pbCanAddDeleteWebParts, Boolean& pbGhostedDocument,
    Boolean& pbDefaultToPersonal, Boolean& pbIsWebWelcomePage, String& pbstrSiteRoot, Guid& pgSiteId, UInt32& pdwVersion, String& pbstrTimeLastModified, String& pbstrContent, UInt32& pdwPartCount, Object& pvarMetaData, Object&
    pvarMultipleMeetingDoclibRootFolders, String& pbst... 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.50* w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Files                          ak8dj High     ...rRedirectUrl, Boolean& pbObjectIsList, Guid&
    pgListId, UInt32& pdwItemId, Int64& pllListFlags, Boolean& pbAccessDenied, Guid& pgDocid, Byte& piLevel, UInt64& ppermMask, Object& pvarBuildDependencySet, UInt32& pdwNumBuildDependencies, Object& pvarBuildDependencies,
    String& pbstrFolderUrl, String& pbstrContentTypeOrder, Guid& pgDocScopeId)     at Microsoft.SharePoint.Library.SPRequest.GetFileAndMetaInfo(String bstrUrl, Byte bPageView, Byte bPageMode, Byte bGetBuildDependencySet, String
    bstrCurrentFolderUrl, Int32 iRequestVersion, Byte bMainFileRequest, Boolean& pbCanCustomizePages, Boolean& pbCanPersonalizeWebParts, Boolean& pbCanAddDeleteWebParts, Boolean& pbGhostedDocument, Boolean& pbDefaultToPersonal, Boolean&
    pbIsWebWelcomePage, String& pbstrSiteRoot, Guid& pgSiteId, UInt32& pdwVersion,... 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.50* w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Files                          ak8dj High     ... String& pbstrTimeLastModified, String& pbstrContent,
    UInt32& pdwPartCount, Object& pvarMetaData, Object& pvarMultipleMeetingDoclibRootFolders, String& pbstrRedirectUrl, Boolean& pbObjectIsList, Guid& pgListId, UInt32& pdwItemId, Int64& pllListFlags, Boolean& pbAccessDenied,
    Guid& pgDocid, Byte& piLevel, UInt64& ppermMask, Object& pvarBuildDependencySet, UInt32& pdwNumBuildDependencies, Object& pvarBuildDependencies, String& pbstrFolderUrl, String& pbstrContentTypeOrder, Guid& pgDocScopeId)    
    at Microsoft.SharePoint.SPWeb.GetWebPartPageContent(Uri pageUrl, Int32 pageVersion, PageView requestedView, HttpContext context, Boolean forRender, Boolean includeHidden, Boolean mainFileRequest, Boolean fetchDependencyInformation, Boolean& ghostedPage,
    String& siteRoot, Guid& siteId, Int64& bytes, ... 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.50* w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Files                          ak8dj High     ...Guid& docId, UInt32& docVersion, String&
    timeLastModified, Byte& level, Object& buildDependencySetData, UInt32& dependencyCount, Object& buildDependencies, SPWebPartCollectionInitialState& initialState, Object& oMultipleMeetingDoclibRootFolders, String& redirectUrl, Boolean&
    ObjectIsList, Guid& listId)     at Microsoft.SharePoint.ApplicationRuntime.SPRequestModuleData.FetchWebPartPageInformationForInit(HttpContext context, SPWeb spweb, Boolean mainFileRequest, String path, Boolean impersonate, Boolean&
    isAppWeb, Boolean& fGhostedPage, Guid& docId, UInt32& docVersion, String& timeLastModified, SPFileLevel& spLevel, String& masterPageUrl, String& customMasterPageUrl, String& webUrl, String& siteUrl, Guid& siteId, Object&
    buildDependencySetData, SPWebPartCollectionInitialState& initialState, ... 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.50* w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Files                          ak8dj High     ...String& siteRoot, String& redirectUrl, Object&
    oMultipleMeetingDoclibRootFolders, Boolean& objectIsList, Guid& listId, Int64& bytes)     at Microsoft.SharePoint.ApplicationRuntime.SPRequestModuleData.GetFileForRequest(HttpContext context, SPWeb web, Boolean exclusion, String
    virtualPath)     at Microsoft.SharePoint.ApplicationRuntime.SPRequestModule.InitContextWeb(HttpContext context, SPWeb web)     at Microsoft.SharePoint.WebControls.SPControl.SPWebEnsureSPControl(HttpContext context)    
    at Microsoft.SharePoint.ApplicationRuntime.SPRequestModule.GetContextWeb(HttpContext context)     at Microsoft.SharePoint.ApplicationRuntime.SPRequestModule.PostResolveRequestCacheHandler(Object oSender, EventArgs ea)    
    at System.Web.HttpApplication.SyncEventExecutionStep.System.Web.HttpApplication.IEx... 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.50* w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Files                          ak8dj High     ...ecutionStep.Execute()     at System.Web.HttpApplication.ExecuteStep(IExecutionStep
    step, Boolean& completedSynchronously)     at System.Web.HttpApplication.PipelineStepManager.ResumeSteps(Exception error)     at System.Web.HttpApplication.BeginProcessRequestNotification(HttpContext context, AsyncCallback
    cb)     at System.Web.HttpRuntime.ProcessRequestNotificationPrivate(IIS7WorkerRequest wr, HttpContext context)     at System.Web.Hosting.PipelineRuntime.ProcessRequestNotificationHelper(IntPtr rootedObjectsPointer, IntPtr
    nativeRequestContext, IntPtr moduleData, Int32 flags)     at System.Web.Hosting.PipelineRuntime.ProcessRequestNotification(IntPtr rootedObjectsPointer, IntPtr nativeRequestContext, IntPtr moduleData, Int32 flags)    
    at System.Web.Hosting.UnsafeIISMethods.MgdIndicateCompl... 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.50* w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Files                          ak8dj High     ...etion(IntPtr pHandler, RequestNotificationStatus&
    notificationStatus)     at System.Web.Hosting.UnsafeIISMethods.MgdIndicateCompletion(IntPtr pHandler, RequestNotificationStatus& notificationStatus)     at System.Web.Hosting.PipelineRuntime.ProcessRequestNotificationHelper(IntPtr
    rootedObjectsPointer, IntPtr nativeRequestContext, IntPtr moduleData, Int32 flags)     at System.Web.Hosting.PipelineRuntime.ProcessRequestNotification(IntPtr rootedObjectsPointer, IntPtr nativeRequestContext, IntPtr moduleData, Int32 flags) 
     9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.50  w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Files                          aiv4w Medium   Spent 0 ms to bind 13297 byte file stream 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.52  mssearch.exe (0x1B6C)                    0x3F24 SharePoint Server Search       Crawler:Content
    Plugin         ajjig Medium   CSSFeeder::ReportPingSession: the session 86ba63af-ce5a-4c7c-a836-c70ee9961b91 has been pinged by PollCallbacks 
    07/16/2014 17:32:08.52  w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Information Rights Management  ai4ko Medium   Information Rights Management (IRM): The IRM status of the document [] is: untried (-1), the return values is: 0x0. 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.52  w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Files                          ahjkm Medium   Spent 0 ms to send 13297 byte file stream 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.52  w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Logging Correlation Data       xmnv Medium   Site=/ 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.52  w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Monitoring                     b4ly High     Leaving Monitored Scope (PostResolveRequestCacheHandler). Execution Time=66.9455 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.57  w3wp.exe (0x2CE0)                        0x45CC SharePoint Server            
     Logging Correlation Data       xmnv Medium   Name=Task: SessionManager.PerformOngoingRequestDepartures 6b6b4445-2b72-0002-9138-988147c881a0
    07/16/2014 17:32:08.58  w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     General                        g3ql High     [Forced due to logging gap, cached @ 07/16/2014 17:32:08.55, Original
    Level: Verbose] GetUriScheme(/QAPortal/PackingMaterial/frmFile4) 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.58  w3wp.exe (0x5234)                        0x1E7C SharePoint Foundation        
     Database                       8acb High     [Forced due to logging gap, Original Level: VerboseEx] Reverting to process
    identity 9e13a59c-dffb-b08d-acd9-c283e837f4a5
    07/16/2014 17:32:08.64  OWSTIMER.EXE (0x1ED4)                    0x5B18 SharePoint Foundation          Monitoring                   
     nasq Medium   Entering monitored scope

  • Export and import of 3 Gb table

    Dear ,
    I have table with more than 3 million records, i need to export this table and import into a different database . can you please help me with the export and import options options so that i can speed up the process...
    for you information
    this table hase 3 Fk, 2 Indexs
    i would appreciate your help in this regard
    thanks

    One way to speed up the exp/imp process is to use DIRECT=Y option. Also increase the size of BUFFER parameter.
    While importing large amount of data it is better to drop relational constraints and indexes and recreate them after import.

Maybe you are looking for

  • How can I choose a symbol for presentation a movie in the finder?

    I have several movies composend mainly of single photos. I want to show a movie in the finder with a movie-specific photo. I got the impression that iMovie chooses a phot by random. How can I specify my preferred picture (it should be the very first

  • Help with Zen Micro installat

    I bought a Zen Micro and came home extremely excited, having previously owned a Zen Xtra 30GB complimented by an Audigy 2 soundcard. I put in the installation cd; as the drivers for the Micro are being installed, it stops at 99% saying "Jukebox Drive

  • Creating UDO through code  with 1 Docment table & 4 Document Line Tables

    Hi, I need to create one Document table and four Document Line tables(child tables) and make them UDO through code. This process is working When I am creating UDO with one Document table and one Document Line table but if I add more than one Child ta

  • Budget Entry Message - Restor Account Accumulators

    I have a client running 2007 A.  They went live on January 1, 2010, but have just initialized the budgets and are beginning to enter their budget amount.  During budget entry, they received the message: Account '40100' has transactions in the sytem. 

  • Time machine deletes office license

    Hey everyone, I just got my 1TB LaCie Thunderbolt HDD in the mail last night and set up time machine on it, everything has been working wonderfully except for the little issue where every time a backup is run (either automatically or manually) my Off