IDM export and import commands

Hi,
I am trying to export and import some of the objects in IDM using command line interface.
Eg: I have an user Object which needs to be exported in the form of an XML file using command line and make some changes and import the user object xml back to IDM using a command line interface
I am very new to IDM. Any help regarding this would be useful.
Thanks,
Hsotnas
Edited by: Hsotnas on Sep 18, 2007 7:17 PM

Actually it can. There are a few ways to achieve this but one is.
1. List Objects > ListFile
2. read ListFile to write commands to CommandFile i.e ListFile might contain:
Configuration SPE
EndUserForm MyForm
then this script (perl or whatever) would convert this to the following in CommandFile
"getObject Configuration SPE > SPE.xml"
"getObject EndUserForm MyForm > MyForm.xml"
3. run lh console < CommandFile

Similar Messages

  • Exporting and importing applications/workspaces

    I am currently on a testing workspace and want to put the application/workspace to another workspace. I know the export and import commands but my concern is the application import. The application is linked to a sample table caled "Samp" and now I would need to link it to a table called "test". This table is almost eactly the same as "Samp" witht he exeption of 2 or 3 new columns. I am wondering what I need to do to change the link to the new table and whether the application will still work.
    Thanks

    "419008",
    In cases like this, where I know the table name is going to be different across HTML DB instances, what I usually do is create a view on top of the table, and then build my HTML DB applications against the view. Then, assuming the definition of the tables are the same, it should work without further modification.
    So in your testing workspace you could have:
    create view foo as select * from Samp
    and in your other workspace you could have:
    create view foo as select * from test
    Joel

  • Internal table export and import in ECC 5.0 version

    Hi friends,
    I am trying to export and import internal table from one program to other program.
    The below… export and import commands are not working when I run the program in background (using SUBMIT zxxxx via JOB name NUMBER number…..)
    EXPORT ITAB TO MEMORY id 'ZMATERIAL_CREATE'.
    IMPORT ItAB FROM MEMORY ID 'ZMATERIAL_CREATE'.
    Normally it should work. Since It’s not working I am trying with another alternative..
    i.e EXPORT (ptab) INTERNAL TABLE itab.
    My sap version is ECC 5.0….
    For your information, here I am forwarding sap help. Pls have a look and explain how to declare ptab internal table.
    +Extract from SAP help+
    In the dynamic case the parameter list is specified in an index table ptab with two columns. These columns can have any name and have to be of the type "character". In the first column of ptab, you have to specify the names of the parameters and in the second column the data objects. If the second column is initial, then the name of the parameter in the first column has to match the name of a data object. The data object is then stored under its name in the cluster. If the first column of ptab is initial, an uncatchable exception will be raised.
    Outside of classes you can also use a single-column internal table for parameter_list for the dynamic form. In doing so, all data objects are implicitly stored under their name in the data cluster.
    My internal table having around 45 columns.
    pls help me.
    Thanks in advance
    raghunath

    The export/import should work the way you are using it. Just make sure you are using same memory id and make sure its unique - meaning u are using it only for this itab purpose and not overwriting it with other values. Check itab is not initial before you export in program 1 - then import it in prog2 with same memory id...also check case, I am not sure if its case sensitive...
    Here is how you use the second variant...
    Two fields with two different identifications "P1" and "P2" with the dynamic variant of the cluster definition are written to the ABAP Memory. After execution of the statement IMPORT, the contents of the fields text1 and text2 are interchanged.
    TYPES:
      BEGIN OF tab_type,
        para TYPE string,
        dobj TYPE string,
      END OF tab_type.
    DATA:
      id    TYPE c LENGTH 10 VALUE 'TEXTS',
      text1 TYPE string VALUE `IKE`,
      text2 TYPE string VALUE `TINA`,
      line  TYPE tab_type,
      itab  TYPE STANDARD TABLE OF tab_type.
    line-para = 'P1'.
    line-dobj = 'TEXT1'.
    APPEND line TO itab.
    line-para = 'P2'.
    line-dobj = 'TEXT2'.
    APPEND line TO itab.
    EXPORT (itab)     TO MEMORY ID id.
    IMPORT p1 = text2
           p2 = text1 FROM MEMORY ID id.

  • Export and import of an table data

    Hi All,
    I have a situation where i need to append/import  the production table ( X )  data to test table ( X ) which has some data already in it and should not be lost  during the operation. The record count is around 6 million .
    Any expert suggestion or tips with export and import command is highly appriciable.
    Thanks in advance .

    if you are using
    IMPDP --- use table_exist_action=append
    IMP -- ignore=y
    Lets say
    impdp username/pwd directory=<directory_name> dumpfile=<dumpfile_name> tables=X table_exist_action=APPEND.
    This will append the data to existing table

  • About Export and Import of pages in R12

    Hi,
    I am working on R12 Instance and i need export and import commands of pages for R12.
    I have commands To Import and Export of pages in 11i but not for R12.
    Regards
    Hitesh

    Hi Hitesh;
    those batch script files are placed @ jdevbin\oaext\bin
    or you can use java oracle.jrad.tools.xml.importer.XMLImporter for import
    java oracle.jrad.tools.xml.exporter.XMLExporter for export

  • DHCP Server does not work after Exporting and Importing Using Netsh Command

    Hello Friends :
    I had two dhcp servers in windows server 2003 server , I have upgraded one of them to windows server 2008 32 bit and again i installed a windows server 2008 R2 as an additional Domain Controller , the last scenario was like this :
    srv-1 : windows server 2003 + DHCP = working with no problem
    srv-2 : windows server 2003 + DC + DHCP = Worked without problem
    srv-3 : windows server 2008 R2 + DC = worked without any problems
    I exported the DHCP server configuration on the srv-2 using netsh dhcp server export and Imported them to srv-3 using the
    netsh dhcp server import command , the command completed successfully and i can see all of the scopes without any problems and errors , i have authorized the new server without any problem , all scopes are activated without any problem so i
    disabled the srv-2 DHCP service and unauthorized it from active directory , the problem is that the new server semms that does not lease any address to clients !!!
    1- I have authorized it
    2- I used Rogue Checker tool in client computers they see authorized server without any problems
    3- The same tool in workgroup only shows srv-1 as the DHCP server and does not see other DHCP servers
    4- Bindings are OK and DHCP servers only have one NIC installed on them
    What can i do to make sure my srv-3 DHCP server will work on the network ?
    thanks ...
    MIMO

    Are the clients on another network so you need to configure a DHCP relay agent?
    If you load up perfmon on the dhcp server and remove all counters and then add DHCP counters. Do you see any dhcprequests when you reboot your dhcp clients? This will determine if your server actually receives any dhcp requests.
    Have you check event viewer for any warnings or errors?
    And the classic one restarted the dhcp server service (or reboot)?
    Regards Per-Torben Sørensen http://pertorben.wordpress.com/

  • How to export and import database from command line?

    I am trying to export my old database into new server.
    However, I want to export and import that database from command line.
    Is it possible?
    Thanks.
    SZ

    I used this command
    exp username/password@database full=y
    my export log file is as follwing:
    Connected to: Oracle8i Enterprise Edition Release 8.1.5.0.0 - Production
    With the Partitioning and Java options
    PL/SQL Release 8.1.5.0.0 - Production
    EXP-00037: Export views not compatible with database version
    EXP-00000: Export terminated unsuccessfully

  • I am using Iphoto 11 ver 9.4.3 on mac using oxs 10.8.5 i want to export calendar projects to an external hard drive. what is the easiest way to do this? i have tried export and import but it didn't seem to work.

    I am using Iphoto 11 ver 9.4.3 on mac using oxs 10.8.5 i want to export calendar projects to an external hard drive. my goal is to store them in an external hard drive so it doesn't use up memory on the mac hard drive. is it possible to copy the specific projects without copying the entire library? what is the easiest way to do this? i have tried export and import but it didn't seem to work.

    What do you not understand?
    You can duplicate the iPhoto library (command - D ) and delete everything except the project and its photos from the copy and move that
    Or
    However the calendar takes very little space - it is simpy database entries - it is the photos in the calendar that take space - and for most people you would wnat to keep those photos in your library
    you can use a photo in 50 calendars and it still is only one photo in your library - as I explained calenders do not exist as such - they are simply database entries telling iPhotop how to display the calendar - they take almost no space at all
    LN

  • Require help in understanding exporting and importing statistics.

    Hi all,
    I am bit new to this statistics.
    Can anyone please explain me in detail about these commands.
    1) exec DBMS_STATS.GATHER_TABLE_STATS (ownname => 'MRP' , tabname => 'MRP_ATP_DETAILS_TEMP', estimate_percent => 100 ,cascade => TRUE);
    2) exec DBMS_STATS.CREATE_STAT_TABLE ( ownname => 'MRP', stattab => 'MRP_ATP_3');
    3) exec DBMS_STATS.EXPORT_TABLE_STATS ( ownname => 'MRP', stattab => 'MRP_ATP_3', tabname => 'MRP_ATP_DETAILS_TEMP',statid => 'MRP27jan14');
    4) exec DBMS_STATS.IMPORT_TABLE_STATS ( ownname => 'MRP', stattab => 'MRP_ATP_3', tabname => 'MRP_ATP_DETAILS_TEMP');
    I understand that these commands are used to export and import table statistics.
    But please anyone help me in understanding this indetail.
    Thanks in advance.
    Regards,
    Shiva.

    Shiva,
    Please post the details of the application release, database version and OS.
    Please see (FAQ: Statistics Gathering Frequently Asked Questions (Doc ID 1501712.1) -- What is the difference between DBMS_STATS and FND_STATS).
    For exporting/importing statistics summary of FND_STATS Subprograms can be found in (Doc ID 122371.1)
    http://etrm.oracle.com/pls/et1211d9/etrm_pnav.show_details?c_name=FND_STATS&c_owner=APPS&c_type=PACKAGE&c_detail_type=source
    http://etrm.oracle.com/pls/et1211d9/etrm_pnav.show_details?c_name=FND_STATS&c_owner=APPS&c_type=PACKAGE%20BODY&c_detail_type=source
    Thanks,
    Hussein

  • Client Export and Import

    Hi,
    I am having one doubt regarding Client export and import
    In my organization we are having three system landscape, but the Quality and Production server(In Process)
    Development box is ready to use, and we are having another sand server in different location for our consultant practice
    that system is not in our landscape and that is different location,i want to import that client in to my development server hear
    It is possible to export that data from that Client(scc8) copy that in my development server and Import(scc7) is possible..?
    if possible than what are the pre-requisites to do That Export and Import of client between two different systems,
    what is best procedure to do this Task.......!
    Regards'
    Mohammed.Khan

    >>once export is done successfully than can i copy those exported file in to my file server and then copy to my development >>system trans directory (cofiile,datafile) than can i import those files in my development server with *stms_import* command
    Yes, you have to
    >>than can i go for client import(scc7)
    Yes.
    >>***Configuring the TMS between that sand box and my development server***
    Not required but yes if the sandbox server is not in the landscape you have to create a Virtual System in sandbox and you have to configure the Transport, then only you would be able to create the export file
    Regards
    Subhash

  • Export and import of data not table and data ????

    hii brothers and sister
    plz i have a a quetion about Export and import of data in oracle forms
    i have created 02 boutons one for export his trigger like this :
    eclare
    alrt number;
    v_directory varchar2(200) := 'c:\backup'; --- that if the C Drive not the Drive that the windows had installed in it.
    path varchar2(100):='back_up'
    ||to_char(sysdate,'dd_mm_yyyy-hh24_mi_ss');
    v_exp varchar2(200) := 'exp hamada/hamada2013@orcl file = '
    ||v_directory
    ||'\'
    ||path
    ||'.dmp';
    begin
    host(v_exp);
    alrt:=show_alert('MSG');
    end;
    and the bouton import is like this
    declare
    alrt number ;
    v_ixp varchar2(200) := 'imp userid=hamada/hamada2013@orcl file =c:\backup2\back.dmp full=yes';
    begin
    host(v_ixp);
    alrt:=show_alert('MSG');
    ref_list;
    end;
    i have just one table "phone"
    this code is correct he expot not only the data but also the creation of the table ....for exemple i do export and evry thing is good until now and i find the .dmp in the folder backup .. but when i deleted all data from my app and try to import this .dmp iit show me error it tell me thet the table phone is already created ....
    so plz help wanna just export the data of phone not the creation of table and data ???? or how can i import just the data from this .dmp ??

    Pl post OS and database versions.
    You will likely need to use the IGNORE flag - http://docs.oracle.com/cd/E11882_01/server.112/e22490/original_import.htm#sthref2201
    Imp utility (by default) will try and create the table first. Since the table already exists, imp reports an error. Use the IGNORE flag in your imp command to not report such errors.
    HTH
    Srini

  • Export and Import of mappings/process flows etc

    Hi,
    We have a single repository with multiple projects for DEV/UAT and PROD of the same logical project. This is a nightmare for controlling releases to PROD and in fact we have a corrupt repository as a result I suspect. I plan to split the repository into 3 separate databases so that we have a design repos for DEV/UAT and PROD. Controlling code migrations between these I plan to use the metadata export and subsequent import into UAT and then PROD once tested. I have used this successfully before on a project but am worried about inherent bugs with metadata export/imports (been bitten before with Oracle Portal). So can anyone advise what pitfalls there may be with this approach, and in particular if anyone has experienced loss of metadata between export and import. We have a complex warehouse with hundreds of mappings, process flows, sqlldr flatfile loads etc. I have experienced process flow imports that seem to lose their links to the mappings they encapsulate.
    Thanks for any comments,
    Brandon

    This should do the trick for you as it looks for "PARALLEL" therefore it only removes the APPEND PARALLEL Hint and leaves other Hints as is....
    #set current location
    set path "C:/TMP"
    # Project parameters
    set root "/MY_PROJECT"
    set one_Module "MY_MODULE"
    set object "MAPPINGS"
    set path "C:/TMP
    # OMBPLUS and tcl related parameters
    set action "remove_parallel"
    set datetime [clock format [clock seconds] -format %Y%m%d_%H%M%S]
    set timestamp [clock format [clock seconds] -format %Y%m%d-%H:%M:%S]
    set ext ".log"
    set sep "_"
    set ombplus "OMBplus"
    set omblogname $path/$one_Module$sep$object$sep$datetime$sep$ombplus$ext
    set OMBLOG $omblogname
    set logname $path/$one_Module$sep$object$sep$datetime$ext
    set log_file [open $logname w]
    set word "PARALLEL"
    set i 0
    #Connect to OWB Repository
    OMBCONNECT .... your connect tring
    #Ignores errors that occur in any command that is part of a script and moves to the next command in the script.
    set OMBCONTINUE_ON_ERROR ON
    OMBCC "'$root/$one_Module'";      
    #Searching Mappings for Parallel in source View operators
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Searching for Loading/Extraction Operators set at Parallel";
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Searching for Loading/Extraction Operators set at Parallel";
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
    foreach mapName [OMBLIST MAPPINGS] {
    foreach opName [OMBRETRIEVE MAPPING '$mapName' GET TABLE OPERATORS] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT)] {
    if { [ regexp $word $prop1] == 1 } {
    incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET DIMENSION OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET CUBE OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET VIEW OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    if { $i == 0 } {
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Not found any Loading/Extraction Operators set at Parallel";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Not found any Loading/Extraction Operators set at Parallel";
         } else {
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Fixed $i Loading/Extraction Operators set at Parallel";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Fixed $i Loading/Extraction Operators set at Parallel";
    close $log_file;
    Enjoy!
    Michel

  • Releases of export and import systems have to be identical

    hi all,
    we want PRD data on DEV , PRD client has been exported successfully. data ( 2 requests from cofile & data) from PRD copied to DEV in cofile and data.
    now when we import these requests in DEV client through STMS , it say that "Releases of export and import systems have to be identical"
    while oracle patch, kernel patch and java version all are same on both systems.
    plz guide
    Regards
    Khan

    Friends,
    We also faced same issue while restoring USER Master backup in one of our system. Even after applying OSS Note 1291394 it didn't work for us. Though the permanent solution is to apply this OSS Note only.
    As a teporary workaround we can add the TR through OS command using Uncoditional Mode "Overwrite Original" and import also using Uncoditional Mode "Overwrite Original". Below are the commands.
    To Add in buffer:
    tp addtobuffer <TR> SID Client=<> U012 pf=<>
    To Import in system:
    tp import <TR> SID U012 pf=<>
    Regards,
    Mukul SHarma

  • Calling oracle export and import utilities in java

    Hi,
    can someone help me, i have to call oracle export and import utilities through my java application, can anyone guide me, how i can do it..
    Thansks
    Ashish

    Hi,
    can someone help me, i have to call oracle export and
    import utilities through my java application, can
    anyone guide me, how i can do it..You mean the command line ones? Use Runtime.exec();

  • Online export and import in DS6

    Hello,
    We are running Sun Java System Directory Server Enterprise Edition 6.0 on Solaris 10.
    We need to export ou=people,o=xyz.com,dc=xyz,dc=com ONLINE and export to other Server (running same version) ONLINE (without stopping the services). So i ran the below:
    ./dsconf export -h webmail.xyz.com -p 389 ou=people,o=xyz.com,dc=xyz,dc=com ppl.ldif
    prompts for "cn=Directory Manager" password and after entering it says "ou=people,o=xyz.com,dc=xyz,dc=com" does NOT exist. But it actually exists.
    We're able to export "dc=xyz,dc=com" without any problems but we need ou=people only.
    Appreciate if anyone could provide solution with exact commands to run for export and import (without stopping the DS)
    Thanks much
    Prvn

    "dsconf export" operates only on suffix DNs (i.e., usually the base of your DIT which corresponds to the LDBM backend database).
    Usage (from the dsconf manpage):
    dsconf export [-h host] [-p port] [-aQ] [-f  FLAG] ... [[-s DN] ... | [-x DN] ...] SUFFIX_DN [SUFFIX_DN...] LDIF_FILE
    Look into these options and see if they can help limit its output:
    -s DN
    --include DN
    Exports all data under specified DN.
    -x DN
    --exclude DN
    Does not import or export data contained under the specified DN.

Maybe you are looking for

  • Authentication error message

    I recently bought a BT 600 wifi extender, since then although signal has improved I am repeatedly getting an " authentication error" on my mobile phone and kindle fire and I have to keep reconnecting to the WiFi, sometimes several times a day. I have

  • Copa derivation rule for customer group

    Hi All, I am currently facing an issue with Derivation rule and hope someone can help me on resolving  this.I have created a COPA derivation rule which has field as Source Field company code Plant Target field Distribution channel Condition customer

  • At-mio-16xe-50 does not work with labview

    I am putting together another system. I set the daq board with the values recorded from the first system using ni-daq. The setup passes the test in ni-daq. I get the correct readings using ni-daq, but when I go to my labview application I do not get

  • Can't update JDeveloper 11.1.1.6.0 with the SOA Composite Editor

    Greetings, I've installed the Jdeveloper 11.1.1.6.0 and want now to update it with the SOA Composite Editor. After having gone to Help->Check for updates, selected "Search update centers" and checked "Oracle Fusion Middleware Products" and "Official

  • Re Download Purchased Songs after hard disc format

    Hello everyone i have a problem with my iTunes on my Macbook Pro...I had format my hard disc and my purchased music is gone... How can i re-download my lost music? Thanks