Export/Import through R3SETUP.EXE

Dear All,
I am currently running on SAP 4.6C SR2 with 4.6D Kernel on SQL Database. The Dtabase size is 2 TB.
I wish to Export/Import the Database through R3Setup.EXE.
Can anybody help me with steps to do the same & other considerations to be kept in mind.
I need to split Database fiels in 20 equal files.

Hi,
check following document  Chapter 8 R3load procedure on windows
https://websmp206.sap-ag.de/~sapdownload/011000358700002949502001E/R3HOM.PDF
regards,
kaushal

Similar Messages

  • EXPORT/IMPORT through FMs in BADI not working

    Hi Experts,
    I have checked several posts and implemented my requirement but it's not working. PFB my requirement:
    We have a BADI in which I need to EXPORT an itab and the same is to be IMPORTed into ZFM. I have created 2 FMs under same FG. Now, in the FM which is being called in BADI i have EXPORT and in ZFM i have IMPORT. Code is given below:
    FM from BADI: ZBADI_FM
      IF lv_flag EQ 'X'.
        EXPORT gt_msg TO MEMORY ID 'CHANGED_DATES'.
      ENDIF.
    Another ZFM under same FG: ZFM
      IMPORT gt_msg FROM MEMORY ID 'CHANGED_DATES'.
    gt_msg is declared in TOP that is common for both FMs. However, IMPORT is not happening. I am not getting the values that I exported. sy-subrc is 4.
    Can anyone help me out? Your help is appreciated.

    Hi,
    I can't believe that it will work:
    EXPORT TO MEMORY
    A data cluster in the ABAP memory is available to all programs within a call sequence, whereby data can be handed over to called programs.
    call sequence
    This sequence is created if you can return from the called program to the calling program at the call of an ABAP-Program with SUBMIT ... AND RETURN or CALL TRANSACTION. For this purpose, the data of the internal session of the caller remains on a stack. The programs of a call sequence have collective access to the ABAP Memory. A call sequence can be left completely using the statement LEAVE TO TRANSACTION.
    Regards,
    Klaus

  • Exporting & Importing Physical & Logical schemas, data servers, agents

    Hi,
    I am using ODI 10g.
    I want to export Physical & Logical schemas, data servers, agents from my ODI test environment and to import them to ODI production environment.
    My requirement is to do this export import through some scripts instead of doing it manually.
    Please guide for this.
    Thanks,
    Divya

    Hi Divya,
    Personally i feel rather than exporting individual components/objects, its suggested to export master repository as such.
    You can make use of ODI utility OdiExportMaster (under <ur package>->Tools->Oracle Data Integrator Objects) for exporting and Import Master Repository wizard (All Programs->Oracle-> Oracle Data Integrator->Repository Management-> Master Repository Import )for importing.
    Thanks,
    Guru.

  • Move our Existing Database 8.1.7.3 into 10gR2 through Export Import Utility

    Hi,
    I want to move our Existing Database 8.1.7.3 into 10gR2 through Export Import Utility in Oracle EBS (FINANCILA) environment.
    My Current Setup:
    OS: SUN SPARC 8
    ORACLE APPS: 11.5.7
    DB Version: 8.1.7.3
    My New Setup:
    OS: SUN SPARC 10
    ORACLE APPS: R12
    DB Version: 10gR2
    I have successfully created Upgraded file R12 file system and installed R12 tech stack.
    Now we want to move data from Current to New Setup by using Export/Import utility.
    Please suggest me steps or meta link note .
    Thanks
    Anup

    What shoul be better for my case:
    Upgrade Paths :
    E-Business Suite Releases Available Paths
    11.5.8 with 9.2.0.x database Path A
    11.5.9 (base or CU1) with 9.2.0.x or 10.1.0.x database Path A
    11.5.9 CU2 with 9.2.0.x or 10.1.0.x database Path A or Path C
    11.5.9 CU2 with 10.2.0.2 database Path D
    11.5.9 CU2 with 10.2.0.3 database Path E
    11.5.9 CU2 with 10.2.0.4 database Path B
    11.5.10 (base or CU1) with 9.2.0.x or 10.1.0.x database Path A
    11.5.10.2 with 9.2.0.x or 10.1.0.x database Path A or Path C
    11.5.10.2 with 10.2.0.2 database Path D
    11.5.10.2 with 10.2.0.3 database Path E
    11.5.10.2 with 10.2.0.4 database Path B
    Regards
    Anup

  • Database refresh of a 10gb database through export-import

    hi ,
    I am using oracle 10g R2 . I need to refresh the development server from the production server .
    The size of the database is 10GB . Can I do it with export-import ?
    Thank you .

    besides these mentioned methods, i.e. 1) exp/imp; 2) data pump; 3) transportable tablespaces; 4) rman duplicate. you can also use manual duplication of the production database which is faster and easier I would say. Below is link, which can be used for detailed steps.
    http://www.dba-oracle.com/oracle_tips_db_copy.htm
    Please note that, if your source database can not be shut down, so you can copy your datafiles by putting tablespaces or database into hot backup mode.

  • System copy using SAPInst(Export Import database Independent prcoess failed

    Hello ,
    I am doing a System copy using SAPInst export/import process .
    Source system : SAP Netweaver'04( BW 3.5 , Kernel : UC 640 ,Patch level 196 )
    Export process fails at Phase 2 -  Database Export at R3load jobs running 1,waiting 0 . Below is the log details
    SAPSDIC.log
    (EXP) INFO:  entry for BAPICONTEN                        in DDNTT is newer than in DDNTT_CONV_UC: 20040211101817 > 20000621155733
    (EXP) INFO:  entry for BAPICONTENT255                    in DDNTT is newer than in DDNTT_CONV_UC: 20040211101817 > 20031127161249
    (EXP) INFO:  entry for BAPICONVRS                        in DDNTT is newer than in DDNTT_CONV_UC: 20040211101817 > 20010131174038
    (EXP) INFO:  entry for BAPICREATORDATA                   in DDNTT is newer than in DDNTT_CONV_UC: 20040211101817 > 20000621155733
    (EXP) INFO:  entry for BAPICRMDH1                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175654 > 20031211120714
    (EXP) INFO:  entry for BAPICRMDH2                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175654 > 20031211120714
    (EXP) INFO:  entry for BAPICRMEXP                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175513 > 20031211120627
    (EXP) INFO:  entry for BAPICRMEXT                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175514 > 20031211120627
    (EXP) INFO:  entry for BAPICRMKEY                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175654 > 20031211120714
    (EXP) INFO:  entry for BAPICRMKEY_T                      in DDNTT is newer than in DDNTT_CONV_UC: 20051229175835 > 20031211120803
    (EXP) INFO:  entry for BAPICRMMSG                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175654 > 20031211120714
    (EXP) INFO:  entry for BAPICRMMSG_T                      in DDNTT is newer than in DDNTT_CONV_UC: 20051229175835 > 20031211120803
    (EXP) INFO:  entry for BAPICRMOBJ                        in DDNTT is newer than in DDNTT_CONV_UC: 20051229175514 > 20031211120628
    (EXP) INFO:  entry for BAPICRMPAREX_T                    in DDNTT is newer than in DDNTT_CONV_UC: 20051229175452 > 20031211120305
    (EXP) INFO: limit reached, 5000 tables in DDNTT are newer than in DDNTT_CONV_UC
    (EXP) INFO: NameTab check finished. Result=2  #20100113131216
    (EXP) INFO: check for inactive NameTab entries: Ok.
    (RSCP) INFO: I18N_NAMETAB_TIMESTAMPS not in env: checks are ON (Note 738858)
    (DB) INFO: disconnected from DB
    D:\usr\sap\B35\SYS\exe\run/R3load.exe: job finished with 1 error(s)
    D:\usr\sap\B35\SYS\exe\run/R3load.exe: END OF LOG: 20100113131216
    ***SAPCLUST.log ****
    (NT)  Warn:  EDIDOC: normal NameTab from 20040211095029 younger than alternate NameTab from 20000621155733!
    (EXP) TABLE: "EDIDOC"
    (NT)  Warn:  PCDCLS: normal NameTab from 20040211095029 younger than alternate NameTab from 20000621155733!
    (EXP) TABLE: "PCDCLS"
    (NT)  Warn:  SFHOA: normal NameTab from 20040211095029 younger than alternate NameTab from 20000621155733!
    (EXP) TABLE: "SFHOA"
    (NT)  Warn:  SFHYT: normal NameTab from 20040211095029 younger than alternate NameTab from 20000621155733!
    (EXP) TABLE: "SFHYT"
    (NT)  Warn:  UMG_TEST_C: normal NameTab from 20040211095029 younger than alternate NameTab from 20031113150115!
    (EXP) TABLE: "UMG_TEST_C"
    myCluster (55.22.Exp): 712: error when retrieving alternate nametab description for physical table UMG_TEST_F.
    myCluster (55.22.Exp): 713: return code received from nametab is 32
    myCluster (55.22.Exp): 299: error when retrieving physical nametab for table UMG_TEST_F.
    (CNV) ERROR: data conversion failed.  rc = 2
    (RSCP) INFO: I18N_NAMETAB_TIMESTAMPS not in env: checks are ON (Note 738858)
    (DB) INFO: disconnected from DB
    D:\usr\sap\B35\SYS\exe\run/R3load.exe: job finished with 1 error(s)
    D:\usr\sap\B35\SYS\exe\run/R3load.exe: END OF LOG: 2010011312563
    Please suggest.
    Thanks & Regards
    Ganesh

    Is your DB unicode?  If so, did you select the unicode flag in sapinst?
    This [thread|System Copy Error while exporting ABAP; might offer some help.
    -Zach

  • ST02 Export / Import buffer and directory entries

    What would cause the number of (Export / Import) directory entries to steadily increase in ST02 until there were none left. I currently have the directory size set to 30,000 and over the course of about 7 days they are all used up and then I have extensive swaps. The swaps are at 1,060,431 after about one month.

    lots of data input through interfaces for example...if the values are too small you have to increase them and checek again until the value is sufficient...
    GreetZ, AH

  • SLD Export / import Technical and Business systems

    Dear all,
    I am new to SLD administration.
    We have task to update the local SLD on the Production PI system PPI ( Target ) with the business systems ZEC, ZSC from the Production Master SLD on Production solution manager system
    PSM ( source ) using Export/Import on the necessary technical and business systems.
    Please confirm whether i am following correct steps :
    Go to SLD Administration on PPI target system
    ALL Instances
    Instances by Class
    Model
    ( which one of this option need to be selected , please let me know  ?? )
    Go to SLD Administration on PSM source system.
    select Export button ->
    Select business systems one by one and perform Export which will create .ZIP file
    at local machine.
    Go to SLD Administration on PPI target system
    select Import button ->
    Select Export ZIP file which created in previous steps from PSM and provide location from desktop
    Start Import button ( only if you get Continue import option )
    Please let me know if above procedure is correct for business systems
    if yes then  what need to be done for Technical system ??...do they require Export/import also ?
    If yes when i go to technical system option i get Export option disabled...but same way when i go to
    business systems option i get Export button enable and able to perform the export.
    Do i need to only Export / Import business systems ?
    Plese help me.
    Regards,
    RR

    Manoj,
    It seems entries are added manually .... should we remove and run Import for the same ...
    or just try to overwrite .... in case of overwrite option ( as we have open OSS message ) SAP has suggested through OSS message that If you end up with problems when trying to overwrite, then the PPI SLD is broken and will have to be re-built because someone made a manual change to it and got the GUIDs out of synch with the master.
    I do not understand if target system PPI SLD gets broken during overwriting operation as mentioned by SAP ... whats is the way / steps / procedure to Re-build SLD .
    In rebuilt - process do i need to Import ( Export -> Export lines -> ALL ) the export backup created before start of this activity ? please let me know SLD rebuilt steps and about Remove - reimport issue.
    Appreciate you time on this. thanks in advance.

  • Export/Import Project still buggy in v2.1.3

    After seeing the update notes for v2.1.3 and noticing references to project importing, I started to get excited that maybe the bugs have been ironed out. Unfortunately my longstanding bug still remains.
    If anyone can think of any workarounds I'd be extremely grateful...
    I have a 2.16GHz Intel Core Duo MacBook Pro with 2GB RAM, running OS X 10.5.6 and now Aperture 2.1.3
    I have a project with many images, arranged into stacks, a pick in each stack with many adjustment and a 62-page book created from these picks. There are also some smart albums and a light-table.
    Now I export the project (i.e. in the Project Inspector I right-click the project and choose Export > Project) to a file on the desktop and rename the existing project in Aperture. I now import the saved .approject file and I find that all my Picks are no longer Picks and my book is therefore no longer made up of my desired images. (I've tried this many times now and ruled out the possibility of a corrupt export file).
    As a result, when I select the Book I get the "Non-Pick Images" dialog with the dreaded "One or more placed items are no longer the pick of a stack." message. "Use Current Pick" ruins the book - I have to go through and work out all the picks again (and also re-pan all the images within the Photo Boxes as this data has also been lost). "Create new Version" appears to preserve my book's original look, but my project now contains a couple of hundred more images than it used to and my other albums and light-table still have incorrect images shown as the Picks.
    Does anybody have any ideas of what I can do to ensure the stack-picks are preserved during the export/import process?
    (By the way, the reason I'm exporting and then re-importing is because I actually want to do the export from my laptop where the project is and then the import on my main work machine where the rest of my Aperture library lives, but that fails for the same reason, so I'm doing the export+import on my laptop for now to reduce the number of variables in the problem.)

    I go with the assumption that you now know how to create transport sets. After having created a transport set containing the objects you want to be exported, go to the administer tab under the main page of portal. There you will find the Export/Import portlet. Here you need to choose your transport set from the first LOV and then click EDIT to choose the "Security Option". Now you export out the stuff. This will export all the user page customizations for you.
    Thanks.

  • Migrate Database- Export/Import

    Hi,
    I need to migrate an Oracle database 9i from Sun Solaris to Linux. The final target database version would be 11g.
    Since this is a 9i database, I see that we have only option of export and database. We have around 15 schemas.
    I have some queries related to it.
    1. If I perform a export with full=y rows=y, Does it export Sys, System Schema objects also?
    2. Can we perform export in Oracle 9i and use datapump import on targert 11g?
    3. What is the ebst approach - a) to perform schema by schema export or b) to perform a full database export with exp / file=xxx.dmp log=xxxx.log full=y?
    Since there is a database version different I dont want to touch sys, system schema objects.
    Appreciate your thoughts.
    Regards
    Cherrish Vaidiyan

    Hi,
    Let me try to answer some of these questions you queried for:
    1. If I perform a export with full=y rows=y, Does it export Sys, System Schema objects also?Export won't export sys objects. For example, there are tables in sys, like obj$ that contain information for other metadata objects, like scott.emp, etc. This is not exported becuase when scott.emp is exported, the data from obj$ is essentially exported that way. When the dumpfile is imported and scott.emp is recreated, the data in sys.obj$ will be restored through the create table statement. As far as the SYSTEM schema is concerned, some objects are exported and some are not. There are tables in system that contain information about queues, jobs, etc. These will probably not make any sense on the target system so those types of tables are excluded from the export job. Other objects make sense to export/import so those are done. This is all figured out in the internals of export/import. Thre are other schemas that are not exproted. Some that I can think of are DMSYS, ORDSYS, etc. This would be for the same reason as SYS.
    2. Can we perform export in Oracle 9i and use datapump import on targert 11g?No, the dumpfiles are formatted differently. If you use exp, then you must use imp. If you use expdp, then you must use impdp. You can do exp on 9i and imp on 11g with the dumfile that was created on 9i.
    3. What is the ebst approach - a) to perform schema by schema export or b) to perform a full database export with exp / file=xxx.dmp log=xxxx.log full=y?This is case by case decision. It depends on what you want. If you want the complete database moved, then I would personally think that a full=y is what you would want to do. If you just did schema exports, then you would never export the tablespaces. This would mean that you would have to create the tablespaces on the source system before you ran imp. There are other objects that are not exported when a schema level export is performed that are exproted when a full is performed. This information can be seen in the utilities guide. Look to see what is exported when in user/schema mode vs full/database mode.
    Since there is a database version different I dont want to touch sys, system schema objects.This is all done for you with the internal workings of exp/imp.
    Dean
    Edited by: Dean Gagne on Jul 29, 2009 8:38 AM

  • EXPORT - IMPORT in BACKGROUND JOB

    Hello ABAP Gurus,
    There are two programs I am using.
    In 1st Program I am Exporting the Data to the ABAP Memory and then after that scheduling the second program in background from 1st Program through
    1 ) JOB_OPEN
    2 ) SUBMIT
    3 ) JOB_CLOSE
    Now I am trying to Import the data that is Exported in the second program which is scheduled in background.
    But I am not able to Import it.
    My question is that - Does EXPORT - IMPORT Statements works in such scenario when background job is scheduled.
    As it is working fine if, I only submits and does not put in the Background  JOB.
    Looking forward for the answer.
    Helpful answers will definately be awarded.
    Thanks in Advance
    Sudhanshu Garg

    Hi Sudhanshu,
    Export/import to memory uses ABAP memory which will not be accessible by the background job.
    No need of creating the structures in dictionary
    You can use the Table INDX for storing your data in the database.
    See the link below for an example.
    http://help.sap.com/saphelp_45b/helpdata/en/34/8e73a36df74873e10000009b38f9b8/content.htm

  • Bad performance on system, export/import buffer many sawps

    Hello,
    I have an ECC 6.0 system on AIX with 6 application servers. There seems to be a performance problem on the system, this issue is being noticed very well when people are trying to save a sale order for example, this operation takes about 10 minutes.
    Sometimes we get short dumps TSV_TNEW_PAGE_ALLOC_FAILED or MEMORY_NO_MORE_PAGING but not very often.
    I am not very good at studying the performance issues, but from what I could see is that there are may swaps on buffer export/import, program and generic key. Also the HitRatio is 88% at  buffer export/import, which I think is pretty low.
    I know that the maximum value accepted of swaps per day is 10000, is that right?
    Can you please advice me what needs to be done in order for these swaps to decrese and hit ratio to increase? And also what else I should do in order to analyse and root cause and the bad performance of the system?
    Many thannks,
    manoliv

    Hi,
    sappfpar determines the minimum and maximum (worst-case) swap space requirements of an R/3 application server. It also checks on shared memory requirements and that the em/initial_size_MB and abap/heap_area_total parameters are correctly set with the following procedure:
    /usr/sap/<SYSTEMNAME>/SYS/exe/run/sappfpar check pf=/usr/sap/<SYSTMENAME>/SYS/profile/<Profile name>
    At the end of the list, the program reports the minimum swap space, maximum heap space, and worst case swap space requirements:
    Additional Swap Space Requirements :
    You will probably need to increase the size of the swap space in hosts in which R/3 application servers run.
    As a rule of thumb, swap space should equal
    3 x Size of Main Storage or at least 1 GB, whichever is larger.
    SAP recommends a swap space of 2-3 GB for optimal performance.
    Determining Current Swap Space Availability: memlimits
    You can find out how much swap space is currently available in your host system with R/3’s memlimits program.
    Here’s how to run memlimits:
    From the UNIX command prompt, run the R/3 memlimits program to check on the size of the available swap space on the host system on which an R/3 application server is to run.
    The application server must be stopped, not running.
    /usr/sap/<SYSTEMNAME>/SYS/exe/run/memlimits | more
    The available swap space is reported in the output line Total available swap space: at the end of the program output. The program also indicates whether this amount of swap space will be adequate and determines the size of the data segments in the system.

  • ST02 Export/Import

    Hi
    The system was in  ZERO ADMINISTRATION in st02 later for the performance issue we increased the buffer parameters like
    rsdb/ntab/entrycount 40000 KB                           
    rsdb/ntab/ftabsize   60000 KB                            
    rsdb/ntab/irbdsize   12000 kB
    abap/buffersize 750000 kB
    rsdb/cua/buffersize 6000  kB
    rsdb/obj/buffersize        12000 kB through rz10
    after that in st02 its showing the Export/import twice times
    Export/import          85.03    12,000      5,012      44.43     2,000                      
    Export/import          85.03    12,000      5,012      44.43     2,000
    free directory entries and % free directory's are showing zero
    regards

    hi manoj
    As i previously increased the export/import in st02
    my query is why its showing twice times in st02 like
                                   hit ratio   allocated    free space     %free space      directory size entries 
    Export/import          85.63 ;   12,000         5,147              45.63 ;                      2,000                          
    Export/import          85.63 ;   12,000         5,147              45.63 ;                      2,000               
    free directory entries and % free directory's are showing zero

  • BI Export Import

    ERROR: 2011-06-12 01:50:01 com.sap.inst.migmon.LoadTask run
    Unloading of 'SAPDODS' export package is interrupted with R3load error.
    Process 'E:\bin\uc\R3load.exe -e SAPDODS.cmd -datacodepage 4103 -l SAPDODS.log -stop_on_error' exited with return code 2.
    For mode details see 'SAPDODS.log' file.
    Standard error output:
    sapparam: sapargv( argc, argv) has not been called.
    sapparam(1c): No Profile used.
    sapparam: SAPSYSTEMNAME neither in Profile nor in Commandline
    ==============SAPDODS.log FILE===============
    (EXP) TABLE: "/BIC/AZSD_O2040" #20110612013025
    (EXP) ERROR: DbSlPrepare/BegRead failed
      rc = 103, table "/BIC/B0000104000"
      (SQL error 208)
      error message returned by DbSl:
    Invalid object name '/BIC/B0000104000'.
    Statement(s) could not be prepared.
    (DB) INFO: disconnected from DB
    E:\bin\uc\R3load.exe: job finished with 1 error(s)
    E:\bin\uc\R3load.exe: END OF LOG: 20110612013026
    ===========================================
    While doing the export import process i am geeting an error , can you please advice what could be the reason
    -Addi

    The indexes of all fact and efact table were not correct. We ran a program "RSDD_MSSQL_CUBEANALYZE" after confirmation from SAP to fix the issue.
    After that the issue was fixed.

  • Homogeneous system copy using export/import on WIn/Ora

    Dear experts,
    i need to know the homogeneous system copy procedure using export/import method for following scenario as we are in the process of Data center migration of SAP systemsfrom location to another location.
    Source system:
    ECC6/BI7/APO(ABAP stack)
    OS:WIndows 2003
    D:B:Oracle 10g
    Target System: New hardware installed on Virtual environment(VMware)
    No change of SID
    ECC6/BI//APO
    OS:WINDOWS2008
    D:B: Oracle 11g
    It would be great if anyone can share the documentu2019s/links (Not SAP homogeneous system copy guide) which you have prepared for the homog. system copy using export/import on Windows/oracle  as i have done the system copy using backup/restore for ABAP stack and export/import for Java stack so far.
    Therefore i am not sure about this procedure/prerequisites/checklist needs to prepare as i have gone through the SAP guide but it is not step by step document.
    Cheers

    There are two approacahes for your  project
    1) First wether to use Oracle11g on target database
    2) Continue with Database Oracle 10.2 on traget database
    If you want to use Oracle11.2 on your target database you have to search the installers which are supported for 11.2 database. If the installers are available then only you can install your target database using Oracle11.2.Other wise t use Oracle10.2.0.4 is supported on Win2008r2. You can use this version of Oracle database to build ur target database.It works fine . Absolutely no problem.
    Later on you can upgrade the Oracle10.2 to 11.2 Onwards.
    If installers are avialble you can use Oracle11.2 also. There is absolutely no problem.
    As per my knowledge installers are available and supported with Oracle11.2.Also check the PAM before arriving on any deciscion.

Maybe you are looking for

  • Is an 'external' static IP address required to set up DNS

    Hi there I'm going to be setting up a OSX Server 10.3 machine as an open directory master, to take advantage of single sign-on capabilities. From what I have read in manuals and in this topic: http://discussions.apple.com/thread.jspa?threadID=977178&

  • Can Open Hub create a fixed length file, not CSV?

    I've been given a requirement to produce a file that is not a CSV file, but a file of a fixed length where all of the fields are always in the same positions. For example, a material "batch" number can be of variable length.  They want it to always a

  • VS2013 Community Edition : How to add multiple projects in one solution?

    I am an experienced .Net developer. I am using Visual Studio since 2005. Yesterday, I just installed "Visual Studio 2013 Community Edition". Suddenly I came to know that I can't add multiple projects to a solution. Even a project does not have soluti

  • Oracle enterprise manager and EBS

    Hello, I wonder know what additional software or configuration to make to have E-Business suite 11i monitored by Oracle enterprise Manager 11.1.0.1.0 BR

  • Audigy 2 problems with spdif out not working and cracking appear

    Hello everyone, I am new in this forum but after searching through it i decided to post as i din't find a problem exactly similar to mine. I have a Creative audigy 2 ZS plugged on my computer with this configuration : Gygabyte P35 DS3, Quad core Q660