Export-Import for LSMW

HI, Please let me know how to export the project-subproject-object from one devlopment server to qulity server, IN qulity server we have to import the same objects. please let me know the procedure
Regards
Pinal

Dear All,
I have got few LSMWs created in my Dev System and also in SND systems.
Earlier I used the Export-Import mechanism for moving them to Q & P.
However I am facing issues now.
When I try to export the LSMW it opens me windows dialog box with C: drive but instead of showing me Save button,it shows me open button for click.
When I press that it does not do anything.If I select some other location and click this open button LSMW gets exported in file with the message mapping and rules exported successfully..
Now when I try to import the LSMW in another server, it does not import it & gives me error 'No importable information can be found in the file.'
Am I missing anything here ?
I have already done import export successfully couple of times in the past but why I am getting this error ?
Also, why cant I see Save button while exporting to hard drive and why it does not get exported to C: drive which appears default ?
Request you to suggest please.
Best Regards,
manOO

Similar Messages

  • Parallel export/Import for exchange directory

    Hi,
    We have executed the homogeneous system copy(Win08/Ora 11) on ECC6 EHP4 and in order to speed up the process i have selected the parallel export/Import option on target system as export on source system is been executing by some other vendor.
    During Import abap phase it's displaying as" waiting -- mins for packahe data in exhange directory" as i have provided the path during input phase of "communication parameters for Import" in Exchange directory as <export directory>/Data.
    is this correct path and i have read some where that <TABART>.SQL  will be generated if on source system the same parallel expot/import option has enabled.And i couldn't find *.SQL files on export directory.
    Does it mean that guy who is executing on source system has not consider this option/not generated the SQl file with SMIGR_CREATE_DDL ?
    NB: i am executing import with SAPINST
    Regards,
    Dheeraj

    Dear Dheeraj,
    Kindly check in the export directory path, whether <package>.SGN files been generating or not .
    Also, we can check in exportmonitor.log.After export of individual pacakge, log says package transfered to network share directory.Even if it not merntions in log, parallel exp/imp not possible
    If it generates, parallel import will happen;If not, it will not.
    Its nothing to do with SQL files.
    Regards
    Ram

  • Best practice metadata export/import  for WebCenter portal

    Hi
    I'm using JDeveloper to build the WebCenter portal. When I let's say register a portlet producer along with my portal, the information about it is stored in the metadata. In case I want to move my portal from the integrated WLS which JDeveloper provides to another WebLogic server with WebCenter installation - steps which I need to do are:
    1) Export the metadata from JDeveloper
    2) Import metadata to my target Weblogic server
    3) deploy the application
    Am I missing something?
    To export/import metadata from JDeveloper to another Weblogic (equipped with WebCenter) server I should use the WLST script, right?
    Thanks

    A little correction: IAS Version is Oracle9i Application Server Release 1 version 1.0.2.2.2

  • Command Line Export/Import for ApEx

    Hi There,
    I was wondering if there is an ApEx Utility that allows the Export and Import of applications/pages/images/style sheets etc from the command line on a Linux box.
    I am using Oracle Application Express v3.0.1
    Thanks.
    Tony.

    Tony,
    Another post in the past has discussed about this already.
    Exporting/Importing a *Bunch* of Applications
    John Scott wrote a blog post about this as well at [http://jes.blogs.shellprompt.net/2006/12/12/backing-up-your-applications/|http://jes.blogs.shellprompt.net/2006/12/12/backing-up-your-applications/]
    In short, you can use APEX utility - APEXExport to export and standard SQL to import.
    Ittichai
    Edited by: ittichai on Mar 16, 2009 9:38 PM

  • Custom directories for export/import for Integration Builder

    Hi,
    XI uses default directories for exporting and importing Integration Builder objects, and there is no option to change the directory or to pick up objects from another directory.
    Does anybody know of any way we can change the default directory that XI uses? Maybe some system setting hidden somewhere deep?
    Thanks and Regards
    Manish

    Hi Manish,
    when there's no way there's always an <b>XI way</b>:)
    you can help your basis team using XI
    set you file adapter to monitor export directory
    and try sending your xi export files to a different server/location
    you can have a look how Felix did similar scenario: 
    /people/sap.user72/blog/2005/07/15/copy-a-file-with-same-filename-using-xi
    isn't this a good way ?
    Regards,
    michal

  • Homogeneous system copy using export/import on WIn/Ora

    Dear experts,
    i need to know the homogeneous system copy procedure using export/import method for following scenario as we are in the process of Data center migration of SAP systemsfrom location to another location.
    Source system:
    ECC6/BI7/APO(ABAP stack)
    OS:WIndows 2003
    D:B:Oracle 10g
    Target System: New hardware installed on Virtual environment(VMware)
    No change of SID
    ECC6/BI//APO
    OS:WINDOWS2008
    D:B: Oracle 11g
    It would be great if anyone can share the documentu2019s/links (Not SAP homogeneous system copy guide) which you have prepared for the homog. system copy using export/import on Windows/oracle  as i have done the system copy using backup/restore for ABAP stack and export/import for Java stack so far.
    Therefore i am not sure about this procedure/prerequisites/checklist needs to prepare as i have gone through the SAP guide but it is not step by step document.
    Cheers

    There are two approacahes for your  project
    1) First wether to use Oracle11g on target database
    2) Continue with Database Oracle 10.2 on traget database
    If you want to use Oracle11.2 on your target database you have to search the installers which are supported for 11.2 database. If the installers are available then only you can install your target database using Oracle11.2.Other wise t use Oracle10.2.0.4 is supported on Win2008r2. You can use this version of Oracle database to build ur target database.It works fine . Absolutely no problem.
    Later on you can upgrade the Oracle10.2 to 11.2 Onwards.
    If installers are avialble you can use Oracle11.2 also. There is absolutely no problem.
    As per my knowledge installers are available and supported with Oracle11.2.Also check the PAM before arriving on any deciscion.

  • Import for runtime sharing issue - CS5.5 AS2.0

    I would like to use export/import for runtime sharing option so I can reuse assets form one swf library in another timeline.
    I found this tutorial that explains how to use it, there is also a download link to see the example working:
    http://www.flashkit.com/tutorials/Tips_And_Techniques/How_to_u-FreshLau-1000/
    I do I think exactly the same as it is there, but it doesn't work. Strange thing is that if you open main.swf from downloaded project, it works, but if you open the main.fla and publish it wothout any changes even, it will stop working.
    I have also discovered this problem with runtime sharing option. I have set up my movie clip runtime sharing option from library assets properties to import, as it showen in the picture below...
    but after I click "Ok" and wait apropriate time to proces it, and when I reopen the properties panel of that movie clip I noticed that it has changed "Import for runtime sharing" to "Export for runtime sharing"it self: In addition it also adds "Export for Action Script" which I definatelly haven't choose before.
    Any ideas anyone?
    I gues this is some sort of a problem of flash newer Publish Settings incompatiobility with Flash MX 2004, in which this tutorial was made. although when you open it, you can see it is already stage is set up as AS2.0 and flash player 7. But I have also noticed, that once you open main.fla it is tagged with * symbol as it has not been saved. Anyway, the same occures when I build it from scratch in CS5.5 AS2.0.
    I do not want to do it in AS3.0 because it apparantelly doesn't support attachMovie command.

    Hi Szymon,
    Our flash team and a few other flash clients have experienced this issue.
    It is detailed here in this thread:
    http://forums.adobe.com/message/4000804
    There is no fix however and Adobe have not commented on the issue (via twitter @adobe_care or on the forum).  I contacted adobe support, paid for a support contract, reported the issue and pointed them to the thread but they wanted a video of it after many frustrating conversations (Did I mention that adobe support is -horrible-).  Our flash team never generate a movie though regrettably, they still experience it/deal with it.  It's frustrating to say the least but I am determined to get Adobe to say something about it.
    Adobe make good products but getting support from them is like getting blood out of a stone.  I wonder how many people experience the issue but haven't reported it?
    Regards,
    M

  • SAP Export/Import via R3load

    Hello,
    SAP recommended a sorted export/import for oracle databases. A unsorted export/import ist very very quicker than a sorted export/import.
    very quicker means sorted = eg. 56 hours and unsorted eg. 12 hours.
    Is there any problems if i take the unsorted export?
    regards,
    stefan

    Hello Stefan,
    >> SAP recommended a sorted export/import for oracle databases
    That fact relies on database internal facts like clustering factor of indexes to the table. You will benefit of a sorted export in cases of primary indexes access and so on... but this topic is much bigger than your question.
    >> very quicker means sorted = eg. 56 hours and unsorted eg. 12 hours.
    Where did you get this values from? You can split big tables or packages ... or define parallel access paths and so on... and it depends on your database configuration (sort areas, PGA, etc..)
    I have done some migrations/exports with distmon... and optimized it manually ... so i got from round about 72 hours (normal export with some split packages) to only 12 hours. I have done only sorted exports.
    >> Is there any problems if i take the unsorted export?
    Some tables must be exported sorted (for example cluster tables) - have a look at sapnote #954268 ... but generally there are no "problems".
    If you have a BW system... some rules are changing...
    @ Markus:
    >> You mean the usage of "R3load -loadprocedure fast"? Or how do you do the "unsorted unload"?
    The option "loadprocedure fast" has nothing to do with the unload process. It speeds up the import (insert) by bypassing the buffer cache. Refer to sapnote #1045847 and the following link:
    http://download-east.oracle.com/docs/cd/B10501_01/server.920/a96524/c21dlins.htm
    Regards
    Stefan

  • LSMW of runs with a mandator export/- import

    At present we migrate our system landscape of SAP the releases 4.6C on mySAPERP2005. Before the migration of the quality assurance system (FQ2) we imported here in the mandator 151 a mandator export of the productive mandator (FP2 system/mandator 150) with the profile SAP_ALL. Shortly thereafter the Upgrade on mySAPERP2005 was then accomplished. The specialized divisions announce now that the defined LSMW of runs differs in the FQ2 system from the FP2 system (production) and/or still the conditions before the mandator import has. It looks apparent in such a way, as if the runs and/or the definition those does not stand behind it at all was copied. Does the LSMW become runs with a mandator export/- import with copies? How can we examine that?

    Hi,
    Are you running the report for all the selections(all the data )?
    execute the report with data selection and check?
    Is the data in the Ic - compressed?
    Regards
    KP

  • Best choice for exporting / importing EUL

    Hi all
    I have been tasked with migrating an EUL from a R11 to a R12 environment. The Discoverer version on both environments is 10.1.2 and the OS is Solaris on oracle db's.
    I am unfortunately not experienced with Discoverer and there seems to be no one available to assist for various reasons. So I have been reading the manual and forum posts and viewing metalink articles.
    I tried exporting the entire eul via the wizard and then importing it to the new environment but i was not successfull and experienced the system hanging for many hours with a white screen and the log file just ended.
    I assumed this was a memory problem or slow network issues causing this delay. Someone suggested I export import the EUL in pieces and this seemed to be effective but I got missing item warnings when trying to open reports. This piece meal approach also worried me regarding consistency.
    So I decided to try do the full import on the server to try negate the first problem I experienced. Due to the clients security policies I am not able to open the source eul and send it to our dev. I was able to get it from their dev 11 system but I dismissed this as the dev reports were not working and the only reliable eul is the Prod one. I managed to get a prod eex file from a client resource but the upload to my server was extremely slow.
    I asked the dba to assit with the third option of exporting a db dump of the eul_us and importing this into my r12 dev environment. I managed this but had to export the db file using sys which alleviated a priviledge problem when logging in, I have reports that run and my user can see the reports but there are reports that were not shared to sysadmin in the source enviroment are now prefixed with the version 11 user_id in my desktop and the user cannot see her reports only the sysadmin ones.
    I refreshed the BA's using a shell script I made up which uses the java cmd with parameters.
    After some re reading I tried selecting all the options in the validate menu and refreshing in the discover admin tool.
    If I validate and refresh the BA using the admin tool I get the hanging screen and a lot of warnings that items are missing( so much for my java cmd refresh!) and now the report will not open and I see the substitute missing item dialogue boxes.
    My question to the forum is which would be the best approach to migrate the entire eul from a R11 instance to a R12 instance in these circumstances?
    Many thanks
    Regards
    Nick

    Hi Srini
    The os and db details are as follows:
    Source:
    eBus 11.5.2
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
    SunOS 5.10 Generic_142900-11 sun4u sparc SUNW,Sun-Fire-V890
    Target:
    ebus 12.1.2
    Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production DEV12
    SunOS 5.10 Generic_142900-11 sun4u sparc SUNW,Sun-Fire-V890
    Yes the DBA initially did an exp for me using EUL_US as the owner but a strange thing happened with privileges and also, some of the imported tables appeared in the target environment under the apps schema(21 tables) even though the eul_us exp had been 48 tables.
    I also had a problem on the db with "eul_us has insufficient privileges on table space discoverer" type errors.
    I checked the eul_us db privileges and was unable to resolve this initial privilege error even though the privileges were granted to eul_us.
    The dba managed to exp as system and then import it with the full=y flag in the import command which seems to bring in the privileges.
    Then I ran the eul5_id.sql and then made up a list of the business areas and made a sh script to refresh the business areas as follows:
    java -jar eulbuilder.jar -connect sysadmin/oracle1@dev -apps_user -apps_responsibility "System Administrator" -refresh_business_area "ABM Activities" -log refresh.log
    This runs successfully and I can log in select business area and grant access to the users. The reports return data.
    Then one of the users said she can't see all her reports. I noticed some if I opened desktop that were sitting there prefixed with a hash and her version 11 user id.
    So back to the manuals and in the disco admin help the instructions are to first go to view > validate > select all options then go to the business area and click file refresh. This gives me a lot of warnings about items that are missing. I assume this is because the item identifiers brought across in the db dump are the version 11 ones and thus not found in the new system.
    Any suggestions?
    Many thanks
    Nick

  • Regarding Distribution Monitor for export/import

    Hi,
    We are planning to migrate the 1.2TB of database from Oracle 10.2g to MaxDB7.7 . We are currently testing the database migration on test system for 1.2TB of data. First we tried with just simple export/import i.e. without distribution monitor we were able to export the database in 16hrs but import was running for more than 88hrs so we aborted the import process. And later we found that we can use distribution monitor and distribute the export/import load on multiple systems so that import will get complete within resonable time. We used 2 application server for export /import but export completed within 14hrs but here again import was running more than 80hrs so we aborted the import process. We also done table splitting for big tables but no luck. And 8 parallel process was running on each servers i.e. one CI and 2 App servers. We followed the document DistributionMonitorUserGuide from SAP. I observerd that  on central system CPU and Memory was utilizing above 94%. But on 2 application server which we added on that servers the  CPU and Memory utilization was very low i.e. 10%. Please find the system configuration as below,
    Central Instance - 8CPU (550Mhz) 32GB RAM
    App Server1 - 8CPU (550Mhz) 16GB RAM
    App Server2 - 8CPU (550Mhz) 16GB RAM
    And also when i used top unix command on APP servers i was able to see only one R3load process to be in run state and all other 7 R3load  process was in sleep state. But on central instance all 8 R3load process was in run state. I think as on APP servers all the 8 R3load process was not running add a time that could be the reason for very slow import.
    Please can someone let me know how to improve the import time. And also if someone has done the database migration from Oracle 10.2g to MaxDB if they can tell how they had done the database migration will be helpful. And also if any specific document availble for database migration from Oracle to MaxDB will be helpful.
    Thanks,
    Narendra

    > And also when i used top unix command on APP servers i was able to see only one R3load process to be in run state and all other 7 R3load  process was in sleep state. But on central instance all 8 R3load process was in run state. I think as on APP servers all the 8 R3load process was not running add a time that could be the reason for very slow import.
    > Please can someone let me know how to improve the import time.
    R3load connects directly to the database and loads the data. The quesiton is here: how is your database configured (in sense of caches and memory)?
    > And also if someone has done the database migration from Oracle 10.2g to MaxDB if they can tell how they had done the database migration will be helpful. And also if any specific document availble for database migration from Oracle to MaxDB will be helpful.
    There are no such documents available since the process of migration to another database is called "heterogeneous system copy". This process requires a certified migration consultant ot be on-site to do/assist the migraiton. Those consultants are trained specially for certain databases and know tips and tricks how to improve the migration time.
    See
    http://service.sap.com/osdbmigration
    --> FAQ
    For MaxDB there's a special service available, see
    Note 715701 - Migration to SAP DB/MaxDB
    Markus

  • Export/Import Process in the UI for Variations Content Translation is Generating CMP Files with No XML

    We have a SharePoint 2010 Publishing Website that uses variations to deliver contain to multiple languages. We are using a third-party translation company to translate publishing pages. The pages are
    exported using the  export/import using the UI process described here: "http://blogs.technet.com/b/stefan_gossner/archive/2011/12/02/sharepoint-variations-the-complete-guide-part-16-translation-support.aspx".
    Certain sub-sites are extremely content-intensive. They may contain many items in the Pages library as well as lists and other sub-sites. 
    For some sub-sites (not all), the exported CMP file contains no XML files. There should be a Manifest.XML, Requirements.XML, ExportSettings.XML, etc., but there are none. After renaming the CMP file
    to CAB and extracting it, the only files it contains are DAT files.
    The only difference I can see between the sub-sites that generate CMP files with no XML files is size. For example, there is one site that is 114 MB that produces a CMP file with no XML files. Small
    sites do not have this problem. If size is the problem, then I would think the process would generate an error instead of creating a single CMP file that contains only DAT files. However, I do not know exactly what the Export/Import Process in the UI is doing.
    This leads to two questions:
    1.
    Does anyone know why some CMP files, when renamed to *.CAB and extracted, would not contain the necessary XML files?
    2. Second, if exporting using the UI will not work, can I use PowerShell? I have tried the Export-SPWeb, but the Manifest.XML does not contain translatable
    content. I have not found any parameters that I can use with Export-SPWeb to cause the exported CMP to be in the same format as the one produced by the Export/Import process in the UI.
    As a next step, we could try developing custom code using the Publishing Service, but before doing this, I would like to understand why the Export/Import process in the UI generates a CMP that
    contains no XML files.
    If no one can answer this question, I would appreciate just some general help on understanding exactly what is happening with the Export/Import Process -- that is, the one that runs when you select
    the export or import option in the Site Manager drop down. Understanding what it is actually doing will help us troubleshoot why there are no XML files in certain export CMPs and assist with determining an alternate approach.
    Thanks in advance
    Kim Ryan, SharePoint Consultant kim.ryan@[no spam]pa-tech.com

    I wanted to bump this post to see about getting some more responses to your problem. I'm running into the same problem as well. We're running a SharePoint 2010 site and are looking at adding variations now. The two subsites with the most content take a
    while to generate the .cmp file (one to two minutes of the browser loading bar spinning waiting on the file). Both files are generated with a lot of .dat files but no .xml files. I was thinking like you that it must be a size issue. Not sure though. Did you
    ever happen to find a solution to this problem?

  • ABAP for Super Dumps: Import- & Export-Parameter for a Table in a FM

    Hello ABAP Profs,
    sorry I am BW.
    <b>Import- & Export-Parameter for a Table in and out of an Function Modul.</b>
    I want to import a table into a Function Module, change it and export it again.
    How do I have to define the Import- and Export- Parameters in the FM ?
    The table looks looks this:
    DATA: zvpshub_tab TYPE SORTED TABLE OF /bic/pzvpshub WITH UNIQUE KEY
    /bic/zvpshub objvers /bic/zvpsoursy INITIAL SIZE 0.
    Thanks a lot
    Martin Sautter

    Hi Clemens,
    <u>in SE11</u> I defined a datatype of Type Structure: ZVPSHUB_ROW.
    <u>in SE11</u> I defiend a datatype of Type Tabletype: ZVPSHUB_TAB,
    bases on Rowtype ZVPSHUB_ROW.
    <u>in SE 80</u> I creates an FM with a CHANGEING Parameter referencing ZVPSHUB_TAB:
    FUNCTION ZVP_SHUB_TAB_LOAD.
    ""Lokale Schnittstelle:
    *"  CHANGING
    *"     VALUE(SHUB_TAB) TYPE  ZVPSHUB_TAB
    <u>in RSA1</u> in BW in the Startroutine of the Upload Rules in defined the table:
    DATA:shub_tab          TYPE zvpshub_tab.
    <u>in RSA1</u> in BW in the Startroutine of the Upload Rules in defined the table:
    DATA:shub_tab          TYPE zvpshub_tab.
    <u>in RSA1</u> in BW in the Startroutine i called the FM
    CALL FUNCTION 'ZVP_SHUB_TAB_LOAD'
        CHANGING
          shub_tab = shub_tab.
    and it works ..
    Thank You
    Martin Sautter

  • I have 100 groups in planning for those 100 groups i want to build roles like interactive,view user,planner etc.for those how to change in export -import folder .xml file  in that edit  how  to change user roles in that xml it will generate automatic id.h

    I have 100 groups in planning for those 100 groups i want to build roles like interactive,view user,planner etc.for those how to change in export -import folder .xml file  in that edit  how  to change user roles in that xml it will generate automatic id.how to do that in xml file ?

    Thanks john for you are reply.
    I had tried what you sad.I open shared service in that foundation project i had export shared service.after that in import-export file.In that role.csv,user.csv,group.csv.Like this file have.When i open user file added some users after i trying save in excel it shown messgse
    I click yes and save the .csv file and import from share servie. i got error like this
    am i doing right way john.or explain clearly

  • Export/import utility for portals: opeasst ; where to launch from?

    hi all
    does this script opeasst.csh have to be run from infra home or the midtier home? i think for exports of pages and portlets from 904, the CH 10 on export/import requires us to launch export from the mid-tier.
    however, if i set up the midtier as my home, and intend to use the opeasst.csh, i do not see this script sitting in the directory where it's been suggested to reside:
    ORACLE_HOME\portal\admin\plsql\wwu
    instead it is in the ORACLE-infra home:
    ORACLE_HOME\portal\admin\plsql\wwu
    please help. it is URGENT.
    I am using OracleAS 10g Portals 9.0.4 version.

    I've been doing some runs and had typical DB export's at about 12 GB per hour (compressed to ~7GB file). Typical DB I/0 bottle neck loosened by using different file system/disk to point export at than where the DB data files reside (CPU ~80% of single CPU)
    now the import is a bigger item.

Maybe you are looking for

  • Need to copy text in e mail and paste to seperate word doc file

    Previous to Firefox 4 I could highlight text in body of e mail, select Edit and drop down to copy then minimize screen and select Word Doc file, then paste and print text. Please advise, Coyote Deb

  • PR price should flow in Contract/sa/po

    hi, i have one requirement that user enter the Price in valuation price in Valuation tab of Purchase requistion . this price should flow into PO and can't be cahnged.Is there any setting is there thanks in advance Rami Reddy

  • Applications not working, permission errors.

    Hi, I have read various threads on here which are asking similar things but I cant seem so find a solution. So heres my situation: I went travelling last year. I have an Imac and before I went I backed it up and transferred the itunes library over to

  • Space between content and logo

    Playing with fluid layouts and getting a space between logo and content DIV's <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd"> <html xmlns="http://www.w3.org/1999/xhtml" xml:lan

  • Problem with email integration..

    Hello to all. I know that this seems to be a common question, but, surprisingly, it is not followed by a common answer too. I got a BB Curve 8520 just 3 days ago. And i'm trying to integrate my gmail account into it, so i can use it's further functio