Export/Import sertificates on same system?

Hi, guru!
Bi 70 & Portal 70 on the same SID (sandbox landscape)
Must I exchange file_s.crt between?
Thanks attn

Hi,
unfortunately noone answers...
Could you please somebody tell me how to do the import via SqlPlus? Maybe it gives more informations about the error...
A.

Similar Messages

  • SLD Export / import Technical and Business systems

    Dear all,
    I am new to SLD administration.
    We have task to update the local SLD on the Production PI system PPI ( Target ) with the business systems ZEC, ZSC from the Production Master SLD on Production solution manager system
    PSM ( source ) using Export/Import on the necessary technical and business systems.
    Please confirm whether i am following correct steps :
    Go to SLD Administration on PPI target system
    ALL Instances
    Instances by Class
    Model
    ( which one of this option need to be selected , please let me know  ?? )
    Go to SLD Administration on PSM source system.
    select Export button ->
    Select business systems one by one and perform Export which will create .ZIP file
    at local machine.
    Go to SLD Administration on PPI target system
    select Import button ->
    Select Export ZIP file which created in previous steps from PSM and provide location from desktop
    Start Import button ( only if you get Continue import option )
    Please let me know if above procedure is correct for business systems
    if yes then  what need to be done for Technical system ??...do they require Export/import also ?
    If yes when i go to technical system option i get Export option disabled...but same way when i go to
    business systems option i get Export button enable and able to perform the export.
    Do i need to only Export / Import business systems ?
    Plese help me.
    Regards,
    RR

    Manoj,
    It seems entries are added manually .... should we remove and run Import for the same ...
    or just try to overwrite .... in case of overwrite option ( as we have open OSS message ) SAP has suggested through OSS message that If you end up with problems when trying to overwrite, then the PPI SLD is broken and will have to be re-built because someone made a manual change to it and got the GUIDs out of synch with the master.
    I do not understand if target system PPI SLD gets broken during overwriting operation as mentioned by SAP ... whats is the way / steps / procedure to Re-build SLD .
    In rebuilt - process do i need to Import ( Export -> Export lines -> ALL ) the export backup created before start of this activity ? please let me know SLD rebuilt steps and about Remove - reimport issue.
    Appreciate you time on this. thanks in advance.

  • Export & Import in the same DataBase

    Hi,
    I'm trying to export and import in the same DataBase but in different User/Tablespace, the imp procces create the objects in the same Tablespace of the export. What can I do to create a copy of a schema in the same database?
    Thanks. JEA

    I am not sure if I am reading your question correctly, Create a new user in your database.
    Use the command line (DOS) command 'exp80 username/password@hoststring'
    there are a number of parameters here, you can find their meanings online if you search for 'export ORACLE command line parameters', I don't know what your requirements are but you have to set the FILE parameter with the name of the file you wish to store your output in(ext should be .dmp).
    To import, use imp80 and log on as the new user you created then the following parameters need to be set:
    File= your file name
    SHOW = n
    IGNORE = n
    GRANTS = y
    ROWS = y
    FROMUSER = old user name
    TOUSER = new user name
    TABLES = *
    I hope this helps.
    SE

  • System copy with Export/Import Distribyted system UNIX database, Windows CI

    HI
    we have a distributed system witch I want to make a copy of
    Oracle Database on HP-UX
    CI + SCS on Windows
    Dual stack system  ABAP/Java
    1)  I have made an system copy with data base copy tools, and that works fine.
    2)  I have made a system copy export/import of a "central system"
    If I try to make an export/import system copy on a distributed system it wont work.
    ( this work fine in a central system) I dont get anny error messages all gren checks marks ! 
    It seems like the problem are in not finding the profiles from CI ?
    It find the profiles for the sapgateway.
    I someone are more interested I have more documentation with screen shots in a word document.
    I Have used Installation Master SR3 and tryed with some older versions as well.
    I have read SAP Note 970518 - Hom./Het.System Copy SAP NetWeaver 7.0 (2004s) SR2
    SAPINST   
    Using virtual hosts for central and/or DB instances
    SYMPTOM:
    You are using virtual host names for your central instance and/or the database instance (e.g. in a HA environment). Although SAPinst finishes successfully, some export steps are not being performed.
    SOLUTION:
    Start the export with the property SAPINST_USE_HOSTNAME=<virtual host>.
    If the central instance and the database instance run on different hosts, run the database export with SAPINST_USE_HOSTNAME=<virtual host name of the DB server> and - if you run a Java or ABAP+Java system - run the central instance export with SAPINST_USE_HOSTNAME=<virtual host nameof the central instance host>.
    If you run a central system with different host names in the profiles, handle the export as a distributed export, calling SAPinst with parameter SAPINST_USE_HOSTNAME. Perform the export of the database instance on the DB server, calling SAPinst with SAPINST_USE_HOSTNAME=<virtual host of the DB server>.

    Hi I have received folloving answer from SAP:
    You're doing everything correct - besides one thing.
    When you specify the profiles during the database export it cannot be
    correct to specify the profiles of the gateway instance located on
    the same host.
    As a consequence - it picks up the wrong instance G01 and sets the
    property 'hasABAP' and 'hasJava' to 'false'...  - thus the export
    is empty. You can see this within the corresponding 'sapinst.log' file.
    The correct procedure would be to specify the profiles of the central
    instance on the Windows Machine. If you cannot access them, you should
    make them available by Samba or by any type of software which makes
    it's possible to export filesystems or directories from windows to
    unix.
    Best regards
    Bent

  • System copy using SAPInst(Export Import database Independent prcoess failed

    Hello ,
    I am doing a System copy using SAPInst export/import process .
    Source system : SAP Netweaver'04( BW 3.5 , Kernel : UC 640 ,Patch level 196 )
    Export process fails at Phase 2 -  Database Export at R3load jobs running 1,waiting 0 . Below is the log details
    SAPSDIC.log
    (EXP) INFO:  entry for BAPICONTEN                        in DDNTT is newer than in DDNTT_CONV_UC: 20040211101817 > 20000621155733
    (EXP) INFO:  entry for BAPICONTENT255                    in DDNTT is newer than in DDNTT_CONV_UC: 20040211101817 > 20031127161249
    (EXP) INFO:  entry for BAPICONVRS                        in DDNTT is newer than in DDNTT_CONV_UC: 20040211101817 > 20010131174038
    (EXP) INFO:  entry for BAPICREATORDATA                   in DDNTT is newer than in DDNTT_CONV_UC: 20040211101817 > 20000621155733
    (EXP) INFO:  entry for BAPICRMDH1                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175654 > 20031211120714
    (EXP) INFO:  entry for BAPICRMDH2                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175654 > 20031211120714
    (EXP) INFO:  entry for BAPICRMEXP                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175513 > 20031211120627
    (EXP) INFO:  entry for BAPICRMEXT                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175514 > 20031211120627
    (EXP) INFO:  entry for BAPICRMKEY                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175654 > 20031211120714
    (EXP) INFO:  entry for BAPICRMKEY_T                      in DDNTT is newer than in DDNTT_CONV_UC: 20051229175835 > 20031211120803
    (EXP) INFO:  entry for BAPICRMMSG                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175654 > 20031211120714
    (EXP) INFO:  entry for BAPICRMMSG_T                      in DDNTT is newer than in DDNTT_CONV_UC: 20051229175835 > 20031211120803
    (EXP) INFO:  entry for BAPICRMOBJ                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175514 > 20031211120628
    (EXP) INFO:  entry for BAPICRMPAREX_T                    in DDNTT is newer than in DDNTT_CONV_UC: 20051229175452 > 20031211120305
    (EXP) INFO: limit reached, 5000 tables in DDNTT are newer than in DDNTT_CONV_UC
    (EXP) INFO: NameTab check finished. Result=2  #20100113131216
    (EXP) INFO: check for inactive NameTab entries: Ok.
    (RSCP) INFO: I18N_NAMETAB_TIMESTAMPS not in env: checks are ON (Note 738858)
    (DB) INFO: disconnected from DB
    D:\usr\sap\B35\SYS\exe\run/R3load.exe: job finished with 1 error(s)
    D:\usr\sap\B35\SYS\exe\run/R3load.exe: END OF LOG: 20100113131216
    ***SAPCLUST.log ****
    (NT)  Warn:  EDIDOC: normal NameTab from 20040211095029 younger than alternate NameTab from 20000621155733!
    (EXP) TABLE: "EDIDOC"
    (NT)  Warn:  PCDCLS: normal NameTab from 20040211095029 younger than alternate NameTab from 20000621155733!
    (EXP) TABLE: "PCDCLS"
    (NT)  Warn:  SFHOA: normal NameTab from 20040211095029 younger than alternate NameTab from 20000621155733!
    (EXP) TABLE: "SFHOA"
    (NT)  Warn:  SFHYT: normal NameTab from 20040211095029 younger than alternate NameTab from 20000621155733!
    (EXP) TABLE: "SFHYT"
    (NT)  Warn:  UMG_TEST_C: normal NameTab from 20040211095029 younger than alternate NameTab from 20031113150115!
    (EXP) TABLE: "UMG_TEST_C"
    myCluster (55.22.Exp): 712: error when retrieving alternate nametab description for physical table UMG_TEST_F.
    myCluster (55.22.Exp): 713: return code received from nametab is 32
    myCluster (55.22.Exp): 299: error when retrieving physical nametab for table UMG_TEST_F.
    (CNV) ERROR: data conversion failed.  rc = 2
    (RSCP) INFO: I18N_NAMETAB_TIMESTAMPS not in env: checks are ON (Note 738858)
    (DB) INFO: disconnected from DB
    D:\usr\sap\B35\SYS\exe\run/R3load.exe: job finished with 1 error(s)
    D:\usr\sap\B35\SYS\exe\run/R3load.exe: END OF LOG: 2010011312563
    Please suggest.
    Thanks & Regards
    Ganesh

    Is your DB unicode?  If so, did you select the unicode flag in sapinst?
    This [thread|System Copy Error while exporting ABAP; might offer some help.
    -Zach

  • Parallel export/Import for exchange directory

    Hi,
    We have executed the homogeneous system copy(Win08/Ora 11) on ECC6 EHP4 and in order to speed up the process i have selected the parallel export/Import option on target system as export on source system is been executing by some other vendor.
    During Import abap phase it's displaying as" waiting -- mins for packahe data in exhange directory" as i have provided the path during input phase of "communication parameters for Import" in Exchange directory as <export directory>/Data.
    is this correct path and i have read some where that <TABART>.SQL  will be generated if on source system the same parallel expot/import option has enabled.And i couldn't find *.SQL files on export directory.
    Does it mean that guy who is executing on source system has not consider this option/not generated the SQl file with SMIGR_CREATE_DDL ?
    NB: i am executing import with SAPINST
    Regards,
    Dheeraj

    Dear Dheeraj,
    Kindly check in the export directory path, whether <package>.SGN files been generating or not .
    Also, we can check in exportmonitor.log.After export of individual pacakge, log says package transfered to network share directory.Even if it not merntions in log, parallel exp/imp not possible
    If it generates, parallel import will happen;If not, it will not.
    Its nothing to do with SQL files.
    Regards
    Ram

  • Regarding Distribution Monitor for export/import

    Hi,
    We are planning to migrate the 1.2TB of database from Oracle 10.2g to MaxDB7.7 . We are currently testing the database migration on test system for 1.2TB of data. First we tried with just simple export/import i.e. without distribution monitor we were able to export the database in 16hrs but import was running for more than 88hrs so we aborted the import process. And later we found that we can use distribution monitor and distribute the export/import load on multiple systems so that import will get complete within resonable time. We used 2 application server for export /import but export completed within 14hrs but here again import was running more than 80hrs so we aborted the import process. We also done table splitting for big tables but no luck. And 8 parallel process was running on each servers i.e. one CI and 2 App servers. We followed the document DistributionMonitorUserGuide from SAP. I observerd that  on central system CPU and Memory was utilizing above 94%. But on 2 application server which we added on that servers the  CPU and Memory utilization was very low i.e. 10%. Please find the system configuration as below,
    Central Instance - 8CPU (550Mhz) 32GB RAM
    App Server1 - 8CPU (550Mhz) 16GB RAM
    App Server2 - 8CPU (550Mhz) 16GB RAM
    And also when i used top unix command on APP servers i was able to see only one R3load process to be in run state and all other 7 R3load  process was in sleep state. But on central instance all 8 R3load process was in run state. I think as on APP servers all the 8 R3load process was not running add a time that could be the reason for very slow import.
    Please can someone let me know how to improve the import time. And also if someone has done the database migration from Oracle 10.2g to MaxDB if they can tell how they had done the database migration will be helpful. And also if any specific document availble for database migration from Oracle to MaxDB will be helpful.
    Thanks,
    Narendra

    > And also when i used top unix command on APP servers i was able to see only one R3load process to be in run state and all other 7 R3load  process was in sleep state. But on central instance all 8 R3load process was in run state. I think as on APP servers all the 8 R3load process was not running add a time that could be the reason for very slow import.
    > Please can someone let me know how to improve the import time.
    R3load connects directly to the database and loads the data. The quesiton is here: how is your database configured (in sense of caches and memory)?
    > And also if someone has done the database migration from Oracle 10.2g to MaxDB if they can tell how they had done the database migration will be helpful. And also if any specific document availble for database migration from Oracle to MaxDB will be helpful.
    There are no such documents available since the process of migration to another database is called "heterogeneous system copy". This process requires a certified migration consultant ot be on-site to do/assist the migraiton. Those consultants are trained specially for certain databases and know tips and tricks how to improve the migration time.
    See
    http://service.sap.com/osdbmigration
    --> FAQ
    For MaxDB there's a special service available, see
    Note 715701 - Migration to SAP DB/MaxDB
    Markus

  • How to Export/Import in Windows-Oracle10g-EnterprisesManager

    Hi
    How to Export/Import in Windows operating system / Oracle 10g through Enterprises manager.
    1)Which oracle user has privilege to logon in Enterprises manager for export/import purpose.
    2)After loging on to the Enterprises manager's Maintaince menu it asks Host credential here whether i should give OS administrator name & password or Oracle user name & password.
    by
    balamuralikrishnan.s

    Hi1)Which oracle user has privilege to logon in Enterprises manager for
    export/import purpose.
    Normally owner of the schema has all privilege of export/import data in his own schema.
    2)After loging on to the Enterprises manager's Maintaince menu it asks Host
    credential here whether i should give OS administrator name & password or Oracle user name & password.Give oracle user name & password(ie., owner of the schema and password)

  • Export import mandotary patches

    hi,
    there is a note on metalink :Oracle Support Master Note for Portal Export/Import.
    Doc ID:      Note:263995.1
    it says that there are mandotary patches required for a successfull export - import.
    For example:
    10.1.2.0.2 (**)
    Patch:5064459 : EXPORT/IMPORT CONSOLIDATED PATCH FOR 10.1.2.0.2 - VERSION 1
    Note:
    To install the Export/Import patch you need to install the latest version of opatch released with Patch:2617419.
    Latest opatch version is 1.0.0.0.55. It was re-released on April 25, 2006.
    10.1.2.0.1 (*) (**)
    There are no patches available for this release.
    10.1.2.0.0 (*) (**)
    Patch:4405137 : EXPORT/IMPORT CONSOLIDATED PATCH FOR 10.1.2.0 - VERSION 1
    Note:
    To install the Export/Import patch you need to install the latest version of opatch released with Patch:2617419.
    Latest opatch version is 1.0.0.0.55. It was re-released on April 25, 2006.
    i didnt understand that.
    my question is do i have to apply these patches to make a successfull export-import between the same version.
    assume that i have exactly the same version 10.2.0.0.0 is it mandatory then?

    i cant beleive that no one has done a export or import on version 10.1.2.0.0
    i am just asking if you have to apply the patch for export between same version (10.1.2.0.0)
    please share your experiences about this issue!

  • Nexus 1010 can not erase export-import folder

    Hi Expert,
    After we import the backup vsb. We try to erase the export-import folder. But system show "Device or resource busy". So we can not execute more export or import command. Please reference the below image and help us to fix this issue.
    Ps. The Nexus 1010 version is "4.2.1.SP1.4"

    Hello Meng,
    I was able to recreate this problem in the lab.  It appears it only occurs when importing a secondary VSB.  A 'system switchover' of the N1010s frees the NFS locks.  Not an ideal workaround, but a workaround.
    CSCtz44247 has been opened for this problem.
    Matthew

  • Page Export/Import problems

    I have been trying to migrate my portal 3.0.7 development to another portal 3.0.8 development database and have been having problems with the page export and import routines. I have been following the Oracle Export / Import document and first come across the problem when importing the content area as it tries to import its content area page.
    I have then tried to export / import pages separately and consistently get the following error.
    pageimp.csh s portal30 p portal30 m reuse d pobpage.dmp c Kenny.ogg security
    Start Portal Page Import
    Please Wait...
    declare
    ERROR at line 1:
    ORA-06510: PL/SQL: unhandled user-defined exception
    ORA-06512: at "PORTAL30.WWUTL_POB_TRANSPORT", line 1879
    ORA-01403: no data found
    ORA-06512: at line 6
    Disconnected from Oracle8i Enterprise Edition Release 8.1.7.0.0 - Production
    With the Partitioning option
    JServer Release 8.1.7.0.0 - Production
    Import of Portal Page Complete
    The Export command used was as follows:
    pageexp.csh s portal30 p portal30 n PAGE0 d pobpage.dmp c tools.ogg security
    As I am migrating from 3.0.7 to 3.0.8 I use the 3.0.7 export script and the 3.0.8 import script. For completeness I have also tried using the 3.0.7 and 3.0.8 import and export scripts, and get the same error.
    The version of Portal I am using is the Solaris version.
    When I look at the new site and click on the pages Tab no new pages are loaded as expected. Does anyone have a similar problem or am I overlooking something?
    Any suggestions?
    Thanks
    Garry
    null

    Portal 3.0.8 is now available in both NT and UNIX. So you need to upgrade both your platforms and then use export/import between the same versions. That's easier said than done as I have yet to get the export/import scripts to correctly and fully work on exporting from NT to UNIX Tru64. Originally I ran the scripts on NT and connected to the UNIX DB; however, somewhere I saw that it is more relaible to run the scripts on UNIX. So I am trying that now.
    Good luck!
    (By the way I notice that Larry declared war on complexity. The laughable irony of this is revealed by a look at these discussion boards!)
    null

  • SYSTEM Copy with Data ( SAP Export/Import way)

    Hello ,
    There is requirement at my client site to build SAP system copy from Production System  without copying Data but
    should have Programs/ Structures / Repository/ Tables & views etc.
    Though We have thought of building SAP system with Export/ Import way and then deleting the Client from copied system
    after that running Remote client copy from Source to Target system with SAP_CUST profile
    But I have heard with SAP Export/Import way , We can have SAP data copy skipped and only structure to copy. If there is any way
    of such kind then Please help me in letting know the same
    Thanks
    Deepak Gosain    

    Hi Deepak
    Kindly refer the SCN link difference between the Client copy Export / import & Remote copy method
    Difference between remote client copy and client Import/Export- Remote client copy steps
    BR
    SS

  • Can i do client export import between systems with different patch levels

    we want to do a client export import between two systems. As I understand the patch levels should be the same in both for this to happen..
    But right now I am in a situation where time is a big constrant. The basis patch level is on par. But the rest are not. Can I do  client import successfully at this stage.

    The only thing that may stop it working could be additional functionality between the systems.
    I work on XI and there is a huge difference between SP12 and SP15.  We were able to do the transports between systems, but as Dev was patched before QA, and the transport was reliant on standard settings not being available, it still went though, but we had to transport it again after applying SP15 to QA.
    Again, it probably depends on what System you have running on Netweaver...

  • Difference between client Export/Import and a system refresh

    Hi Experts,
    I would like to know the diffrence between client Export/Import and System refresh.
    Actually i have done client export from our production system and imported it into our quality system, I have asked our functional consultants whether the production data is now available in Quality. They replied me saying whole data of production system is available in Quality system as off date i have copied the export.
    The size of the export was around 8GB, How can this have all the data of Production system which is around 600 GB.
    In this way what is the use of System refresh which is done by database level ( changing the cntrl file etc... ), As i am getting all the data by doing Client Export/Import.
    Thanks & Regards,
    Injamuri

    Hi,
    What is system refresh when and why it is done?
    The system refersh is nothing but the deletion of the client and replacing the data from other client. For example : you have clients 100, 200 and 300. Suppose when you want to refresh the client 100 you remove the client 100 and replace it with 200 0r 300 as per your reqiurement.
    Mostly the refresh of clients will be happen at the time of development stage.
    System Refresh is a simplified term to Client Copy. Client Copy means copying the production client on to the quality to test the real data. As recommend by SAP this need to carried out every 3 months.
    The process to carry out the same is as follows:
    1. Create a client on quality system using txn scc4
    2. Create a RFC between Production system and Quality System (need to create on quality system)
    3. Login to the newly created client using sap* and pass as a password
    4. Txn sccl to start the client copy.
    You can test the client copy by selecting the test run option. (test run will estimate the time taken for the activity).
    Anil

  • Client Copy Export/Import on large system. 1.5 TB

    Hi,
    We have a 3 tier landscape. DEV, TST, PRD.
    Database size in PRD is 1.5 TB (1500 GB)
    Our developers need fresh master and trans data in TST and DEV.
    We are thinking about making a
    PRD --> TST  (1.5 TB) we do a system copy.
    TST --> DEV (1.5 TB) we can't do a system copy. (Company policy says not to touch repository objects, version management what so ever in DEV....)
    1. How long will a client export / import take from TST --> DEV,  lets say we have "good HW" ?  Will it work ?
    2. I've read about TDMS but its toooo expensive.....is there maybe any 3 party tool that is good for this purpose ? Transfering 1.5 TB of data ...
    Br,
    Martin

    Hi,
    >1. How long will a client export / import take from TST --> DEV, lets say we have "good HW" ? Will it >work ?
    In my opinion, a 1.5 TB database is much too big for a client copy.
    It will work but may last a full week and would degrade the production system performance.
    We have 1.3 TB databases and stopped using client copies since the system was 300 GB...
    We only use database restores and renaming for system refresh even for the dev system.
    There is procedure to keep the versionning of sources.
    Regards,
    Olivier

Maybe you are looking for