Moving Data from 10g to 9i

Hello Everyone,
We are using Oracle 10g for one of our applications.
A third party vendor is going to reuse the data. So the vendor is planning to write a java interface to pull the data from our 10g and move it to their 9i database.
If there are any know issues with this type of migrations please let me know.
I will really appreciate your answers and suggestions
Thanks!

Hi,
There won't be any issues from 10g database to 9i database, but you have to use 9i exp/imp binaries inorder to do this activity.
from upper to lower verision migration always you have to use lower verison binary.
otherwise your import fails

Similar Messages

  • Which CKM is used for moving data from Oracle to delimited file ?

    Hi All
    Please let me know Which CKM is used for moving data from Oracle to delimited file ?
    Also is there need of defining each columns before hand in target datastore. Cant ODI take it from the oracle table itself ?

    Addy,
    A CKM is a Check KM which is used to validate data and log errors. It is not going to assist you in data movement. You will need an LKM SQL to File append as answered in another thread.
    Assuming that you have a one to one mapping, to make things simpler you can duplicate the Oracle based model and create a file based model. This will take all the column definitions from the Oracle based model.
    Alternatively, you can also use an ODI tool odiSQLUnload to dump the data to a file
    HTH

  • Moving Data from Normal table to History tables

    Hi All,
    I'm in the process of moving data from normal tables to
    History tables.
    It can be some sort of a procedure which should be a cron job running at night.
    My aim is to move data say 1.5 yrs or 2yrs old data to History tables.
    What aspects i need to check when moving data. And how can i write a procedure for this requirement.
    The schema is same in both the normal table and history table.
    It has to be a procedure based on particular field RCRE_DT.
    If the rcre_dt is above 2 yrs the data needs to be moved to HIS_<table>.
    I have to insert record in to HIS_table and simultaneously delete record from the normal table.
    This is in Production system and the tables are quite big.
    Pls do find enclosed the attached sample schema for Normal table and HIS_<table>.
    If i want to automate this script as a Cron job for similarly other History tables
    how am i to do it in a single procedure assuming the procedure for moving the data is the same procedure.
    Thanks for ur help in advance.
    SQL> DESC PXM_FLT;
    Name Null? Type
    RCRE_USER_ID NOT NULL VARCHAR2(15)
    RCRE_DT NOT NULL DATE
    LCHG_USER_ID VARCHAR2(15)
    LCHG_DT DATE
    AIRLINE_CD NOT NULL VARCHAR2(5)
    REF_ID NOT NULL VARCHAR2(12)
    BATCH_DT NOT NULL DATE
    CPY_NO NOT NULL NUMBER(2)
    ACCRUAL_STATUS NOT NULL VARCHAR2(1)
    FLT_DT NOT NULL DATE
    OPERATING_CARRIER_CD NOT NULL VARCHAR2(3)
    OPERATING_FLT_NO NOT NULL NUMBER(4)
    MKTING_CARRIER_CD VARCHAR2(3)
    MKTING_FLT_NO NUMBER(4)
    BOARD_PT NOT NULL VARCHAR2(5)
    OFF_PT NOT NULL VARCHAR2(5)
    AIR_CD_SHARE_IND VARCHAR2(1)
    UPLOAD_ERR_CD VARCHAR2(5)
    MID_PT1 VARCHAR2(5)
    MID_PT2 VARCHAR2(5)
    MID_PT3 VARCHAR2(5)
    MID_PT4 VARCHAR2(5)
    MID_PT5 VARCHAR2(5)
    PAX_TYPE VARCHAR2(3)
    PAY_PRINCIPLE VARCHAR2(1)
    SQL> DESC HIS_PXM_FLT;
    Name Null? Type
    RCRE_USER_ID NOT NULL VARCHAR2(15)
    RCRE_DT NOT NULL DATE
    LCHG_USER_ID VARCHAR2(15)
    LCHG_DT DATE
    AIRLINE_CD NOT NULL VARCHAR2(5)
    REF_ID NOT NULL VARCHAR2(12)
    BATCH_DT NOT NULL DATE
    CPY_NO NOT NULL NUMBER(2)
    ACCRUAL_STATUS NOT NULL VARCHAR2(1)
    FLT_DT NOT NULL DATE
    OPERATING_CARRIER_CD NOT NULL VARCHAR2(3)
    OPERATING_FLT_NO NOT NULL NUMBER(4)
    MKTING_CARRIER_CD VARCHAR2(3)
    MKTING_FLT_NO NUMBER(4)
    BOARD_PT NOT NULL VARCHAR2(5)
    OFF_PT NOT NULL VARCHAR2(5)
    AIR_CD_SHARE_IND VARCHAR2(1)
    UPLOAD_ERR_CD VARCHAR2(5)
    MID_PT1 VARCHAR2(5)
    MID_PT2 VARCHAR2(5)
    MID_PT3 VARCHAR2(5)
    MID_PT4 VARCHAR2(5)
    MID_PT5 VARCHAR2(5)
    PAX_TYPE VARCHAR2(3)
    PAY_PRINCIPLE VARCHAR2(1)

    Hi All,
    Thanks for ur valuable suggestion.But can u explain me bit more on this as i'm still confused about switching between partitoned tables and temporary table.Suppose if i have a table called PXM_FLT and an correspoding similar table named HIS_PXM_FLT.How can i do the partitioning shd i do the partitioning on the normal table or HIS_PXM_FLT.i do have a date field for me to partition based on range.Can u pls explain why shd i again create a temp.table.What's the purpose.Now the application is designed in such a way that old records have to be moved to HIS_PXM_FLT.can u pls highlight more on this.Your suggestions are greatly appreciated.As i'm relatively new to this partitioning technique i'm bit confused on how it works.But i came to understand Partitioning is a better operation than the normal insert or delte as it is more data intensive as millions of record need to be moved.Thanks for feedback and ur precious time.

  • Moving data from an array to a JTable

    Hi,
    I am facing some problems with handling array data to JTable.
    My objective is moving data from a 10X5 data array to a JTable.
    I am using the following code:
    String columnNames = { "Col 1", "Col 2", "Col 3", "Col 4", "Col 5" };
    for(int j=0; j<10; j++)
    data[ j ] = array1[ j ] + array2[ j ] + array3[ j ] + array4[ j ] + array5[ j ];
    JTable table = new JTable( data, coulmnNames);
    Can you please help me with the error in tha above code.My array1 is a String array and the other four arrays are of Integer data type.
    Thanks!!!

    hi,
    The actual error message is:
    [root@localhost java]# javac GUI.java
    GUI.java:144: incompatible types
    found : int
    required: java.lang.Object
    data [j] [1] = array1[j];
    ^
    GUI.java:145: incompatible types
    found : int
    required: java.lang.Object
    data [j] [2] = array2[j];
    ^
    GUI.java:146: incompatible types
    found : int
    required: java.lang.Object
    data [j] [3] = array3[j];
    ^
    GUI.java:147: incompatible types
    found : int
    required: java.lang.Object
    data [j] [4] = array4[j];
    ^
    4 errors
    My code segment with line no goes as follows:
    private static JTable createTable() {         
           String[] columnNames = { "Col 1", "Col 2", "Col 3", "Col 4"' "Col 5" };
             Object data [ ] [ ] = new Object [10] [5];
       for( int j=0; j<10; j++)
              data [j] [0] = array1[j]; ...................Line:144
              data [j] [1] = array2[j]; ..................Line 145
              data [j] [2] = array3[j]; ....................Line 146
              data [j] [3] = array4[j]; .................Line 147
              data [j] [4] = array5[j];
      final JTable table = new JTable(data, columnNames);
      return table;
    [code/]
    Array 1 is declared as String array and Array 2,3,4,5 are declared as Integer arrays.The error is coming from the Integer arrays only.
    Java version running on my machine is: 1.4.2
    Please help!!!
    Thanks!!!                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

  • Moving data from flat file

    Hi,
    I am moving data from flat file to oracle table. while populating the oracle table if I get any errors in flat file those errors should populate in ODI error table.
    Is this is possible? if yes. Could you please let me know the set up in ODI.

    CKM is the dedicated for checking the constraints while doing transformation. The constraints includes, PK,FK, conditions etc.,
    There are two types/ways of checking the constraints.
    Flow Control: Triggered CKM after data is loaded in to I$.
    Static Control : Triggered CKM after data is loaded in to target table.
    If you opt for any one the above ODI will create E$ and SNP_CHECK_TAB (summary table for logging the errors) and load the error records.
    ODI will also provide you an option of loading the corrected error records by RECYCLE ERROR feature. In this phase/run ODI will load ONLY the error records in to target table (assuming its been corrected/cleaned).
    **how to set up flow control could you please provide steps**
    **Appreciate your help**

  • Error occurred moving data from the transformation RECON_SEQ_GEN: to the transformation FILTRANS.

    Hi,
    Today one my task in the DAC Etl failed and when I opened the DAC log file of it. I saw the error description as below
    Error while contacting Informatica server for getting workflow status for PLP_SRC_WH_AP_Invoice_Reconciliation
    Error Code = 36331:Unknown reason for error code 36331
    Since there is no much information in DAC log file about the failure, I opened the PLP_SRC_WH_AP_Invoice_Reconciliation session log file.
    The error description in the session log file says.
    There is an error in the port [NEXTVAL]: The default value for the port is set to: ERROR(<<Expression Error>> [ERROR]: transformation error
    ... nl:ERROR(u:'transformation error')).
    2014-01-14 08:00:45 : ERROR : (582188 | TRANSF_1_1_1) : (IS | Oracle_BI_DW_Server) : node01_ : TT_11023 : An error occurred converting data in the port [NEXTVAL]. The row was dropped.
    2014-01-14 08:00:45 : ERROR : (582188 | TRANSF_1_1_1) : (IS | Oracle_BI_DW_Server) : node01_ : TT_11021 : An error occurred moving data from the transformation RECON_SEQ_GEN: to the transformation FILTRANS.
    2014-01-14 08:00:45 : ERROR : (582188 | TRANSF_1_1_1) : (IS | Oracle_BI_DW_Server) : node01 : CMN_1086 : RECON_SEQ_GEN: Number of errors exceeded threshold [1].
    2014-01-14 08:00:45 : INFO : (582188 | TRANSF_1_1_1) : (IS | Oracle_BI_DW_Server) : node01_ : CMN_1053 : : Rowdata: ( RowType=-4(error but dropped due to abort) Src Rowid=1186281 Targ Rowid=1186281
      NEXTVAL (ROW_WID:Int:): "(NULL)"
    I am not sure of this error and I have never seen before. Can any one share some ideas on how to tackle this error and make the task execute successfully.
    Many Thanks.

    Hi,
    fist check weather any special characteristics if not
    check in data source under this we have fields tab check the format of a particular field format internal/external/check u choose internal format, if any check routine once
    use Semantic Groups in the DTP.
    Try  it
    Thanku
    lokeeshM
    Edited by: lmedaSAP_BI on Oct 20, 2010 6:44 AM

  • Moving Data from one table to another table

    Hi,
    I've a requirement to move large volumes (few billions) of data from one table to another empty target table. The target has more number of attributes and may not have same attribute name as the source. Please let me know what's the best method using oracle to move data in this case?
    Thanks in Advance,
    Tom

    You should use nologging and parallelism options
    Please refer to Tom's answers:
    http://asktom.oracle.com/pls/asktom/f?p=100:11:0::::P11_QUESTION_ID:164612348068
    http://asktom.oracle.com/pls/asktom/f?p=100:11:0::::P11_QUESTION_ID:6407993912330
    Kamran Agayev A. (10g OCP)
    http://kamranagayev.wordpress.com
    [Step by Step install Oracle on Linux and Automate the installation using Shell Script |http://kamranagayev.wordpress.com/2009/05/01/step-by-step-installing-oracle-database-10g-release-2-on-linux-centos-and-automate-the-installation-using-linux-shell-script/]

  • Moving data from internal table to Dynamic internal table

    Hello All,
      I am facing a problem i moving the data from a internal table data to Dyn. internal table.
    My code is as follows
    *******Declartion**************************************
    DATA: IT_FIELDCAT TYPE LVC_T_FCAT,
          IS_FIELDCAT LIKE LINE OF IT_FIELDCAT.
    DATA: IT_OUTTAB TYPE STANDARD TABLE OF ZVTEST.
    DATA: WA_OUTTAB TYPE  ZVTEST.
    DATA: LV_LINE LIKE SY-TABIX,
           COUNT TYPE I.
    FIELD-SYMBOLS: <NEW_TABLE> TYPE REF TO DATA.
    FIELD-SYMBOLS: <L_TABLE> TYPE ANY TABLE,
                   <L_LINE> TYPE ANY.
    FIELD-SYMBOLS: <L_FIELD> TYPE ANY.
    FIELD-SYMBOLS: <FS_DATA> TYPE REF TO DATA,
                   <FS_1> TYPE ANY TABLE,
                   <FS_2>,
                   <FS_3>.
    DATA: NEW_LINE TYPE REF TO DATA.
    DATA: LT_DATA TYPE REF TO DATA.
      ASSIGN LT_DATA TO <FS_DATA>.
    Create a new Table
      CALL METHOD CL_ALV_TABLE_CREATE=>CREATE_DYNAMIC_TABLE
      EXPORTING
            IT_FIELDCATALOG = IT_FIELDCAT
      IMPORTING
            EP_TABLE = <FS_DATA>
      EXCEPTIONS
          GENERATE_SUBPOOL_DIR_FULL = 1
          OTHERS                    = 2.
      IF SY-SUBRC = 0.
        ASSIGN <FS_DATA>->* TO <FS_1>.
        CREATE DATA NEW_LINE LIKE LINE OF <FS_1>.
    A field-symbol to access that work area
        ASSIGN NEW_LINE->*  TO <FS_2>.
      LOOP AT IT_OUTTAB INTO WA_OUTTAB.
        MOVE-CORRESPONDING WA_OUTTAB TO <FS_1>.
      ENDLoop.
    But when I am using MOVE-CORRESPONDING WA_OUTTAB TO <FS_1>.
    I am getting the error like
    "<FS_1>" is not a structure or internal table with header line.          
    Please check the code and tell where I am making the mistake.
    Regards,
    Vasanth

    Hello Rich ,
    My is code is as below..
    Declaration:
    DATA: IT_FIELDCAT TYPE LVC_T_FCAT,
          IS_FIELDCAT LIKE LINE OF IT_FIELDCAT.
    DATA: BEGIN OF IT_OUTTAB OCCURS 0,
            PSPID     LIKE PROJ-PSPID, "Projektdefinition
            POSID     LIKE PRPS-POSID, "PSP-Element
            POST1     LIKE PRPS-POST1, "Kurzbeschreibung
            STTXT_INT LIKE CNJ_STAT-STTXT_INT, "Systemstatus
            STTXT_EXT LIKE CNJ_STAT-STTXT_EXT, "Anwenderstatus
          END OF IT_OUTTAB.
    DATA: WA_OUTTAB LIKE IT_OUTTAB.
    DATA: LV_LINES LIKE SY-TABIX,
           COUNT TYPE I.
    FIELD-SYMBOLS:  type any.
    DATA: NEW_LINE TYPE REF TO DATA.
    DATA: LT_DATA TYPE REF TO DATA.
      G_R_DISP_VARIANT-REPORT = SY-REPID.
      G_R_DISP_VARIANT-VARIANT = PA_VAR.
      CALL FUNCTION 'REUSE_ALV_VARIANT_SELECT'
           EXPORTING
                I_DIALOG            = 'N'
                I_USER_SPECIFIC     = 'A'
                I_DEFAULT           = 'X'
                IT_DEFAULT_FIELDCAT = G_T_FIELDCAT
                I_LAYOUT            = G_R_LAYOUT
           IMPORTING
                ET_FIELDCAT         = G_T_FIELDCAT
               et_sort             = l_tab_sort
               et_filter           = l_tab_filter
               ES_LAYOUT           = G_R_LAYOUT
           CHANGING
                CS_VARIANT          = G_R_DISP_VARIANT
           EXCEPTIONS
                ERROR_MESSAGE       = 4
                OTHERS              = 4.
      LOOP AT G_T_FIELDCAT INTO G_R_FIELDCAT WHERE NO_OUT IS INITIAL
                                               AND TECH IS INITIAL.
        MOVE-CORRESPONDING G_R_FIELDCAT TO IS_FIELDCAT.
        IS_FIELDCAT-FIELDNAME = G_R_FIELDCAT-FIELDNAME.
        IS_FIELDCAT-INTTYPE   = G_R_FIELDCAT-INTTYPE.
        IS_FIELDCAT-OUTPUTLEN = G_R_FIELDCAT-OUTPUTLEN.
        IS_FIELDCAT-REF_FIELD = G_R_FIELDCAT-FIELDNAME.
        IS_FIELDCAT-REF_TABLE = G_R_FIELDCAT-REF_TABNAME.
        APPEND IS_FIELDCAT TO IT_FIELDCAT.
      ENDLOOP.
      DATA: WA_LISTE LIKE P_T_LISTE.
      CLEAR:IT_OUTTAB.
      REFRESH: IT_OUTTAB.
      LOOP AT P_T_LISTE INTO WA_LISTE.
        WA_OUTTAB-PSPID = WA_LISTE-PSPID .
        WA_OUTTAB-POSID = WA_LISTE-POSID.
        WA_OUTTAB-POST1 = WA_LISTE-POST1.
        WA_OUTTAB-STTXT_INT = WA_LISTE-STTXT_INT.
        WA_OUTTAB-STTXT_EXT = WA_LISTE-STTXT_EXT.
        APPEND WA_OUTTAB TO IT_OUTTAB.
      ENDLOOP.
    **dynamic table creation for data
      ASSIGN LT_DATA TO .
    Create a new Table
      CALL METHOD CL_ALV_TABLE_CREATE=>CREATE_DYNAMIC_TABLE
      EXPORTING
            IT_FIELDCATALOG = IT_FIELDCAT
      IMPORTING
            EP_TABLE = .
    A field-symbol to access that work area
        ASSIGN NEW_LINE->*  TO  = WA_OUTTAB-PSPID.
         ASSIGN COMPONENT 'POSID' OF STRUCTURE WA_OUTTAB TO <FS_3>.
          ASSIGN COMPONENT 'POSID' OF STRUCTURE  = WA_OUTTAB-POSID.
         ASSIGN COMPONENT 'POST1' OF STRUCTURE WA_OUTTAB TO <FS_3>.
          ASSIGN COMPONENT 'POST1' OF STRUCTURE .
         <FS_5> = <FS_3>.
           = WA_OUTTAB-POST1.
         ASSIGN COMPONENT 'STTXT_INT' OF STRUCTURE WA_OUTTAB TO <FS_3>.
          ASSIGN COMPONENT 'STTXT_INT' OF STRUCTURE  TO P_FILE.
        ENDLOOP.
      ELSE.
        MESSAGE E041(S9) WITH P_FILE.
      ENDIF.
      CLOSE DATASET P_FILE.
    Here my problem is ´´, I am not able to use move corresponding statement.
    So I am not getting the exact values to the fields of the dynamic table...
    Please check it and give me a solution,
    Regards,
    Vasanth

  • Import data from 10g to oracle 9i

    hi,
    how can i import data fom 10g to 9i R2?

    Follow this thread :
    exp/imp between 10g and 9i
    Essentially, you must
    1. Install a 9i server or administrative client environment.
    2. Configure SQLNet connectivity (tnsnames.ora) to the 10g database.
    3. Run exp from the 9i environment, connected to the 10g database.
    4. Create a 9i export dump file.
    5. Import this into your 9i target database.
    If you have different charactersets, you must be mindful of the setting of NLS_LANG in the environment that you run "exp" and "imp" in.

  • What are required Oracle products for moving data from IBM IMS/DB(mainframe) to Oracle environment?

    I am z/OS system programmer, our company is using IMS as its main OLTP database. We are investigating moving data off the mainframe for data warehousing and online fraud detection. One option is using IBM InfoSphere CDC and DB2, another option is using IMS connect and writing our own program, I am wondering what is the oracle solution for this kind of issue?
    I am not oracle technician but I googled and find out Oracle has some product like Oracle Legacy Adapters, OracleAS CDC Adapter and Oracle Connect on z/OS, however I didn't find them in Oracle site(https://edelivery.oracle.com/), I don't know whether these products are deprecated or not?!
    I would very much appreciate any help or guidance you are able to give me

    Thank you for responding.
    I've considered dumping the data into a flat file and using SQL*Loader to import as you suggest but this would require some scripting on a per-table basis. Again: all I want to do is to copy the contents of a table from one database to another. I do not think I should have to resort to creating my own dump and load scripts in order to do that. However I agree with you that this type of solution may be my final solution.
    I've tried the db link solution. It was just a slow as the 'imp' solution for some reason. Don't know why. The tables are rather large (3 tables of a few Gb each) and therefore require intermediate commits when loaded. Otherwise the rollback segment will run out of space. So the 'db link solution' is really a PL/SQL script with a commit for each x records.
    I think Oracle is making it a bit difficult for me to copy the contents of a table from one database to another and to do it efficiently. Perhaps I'm missing something here?

  • Problem while Migrating user data from 10g to 11gR2

    Hi experts,
    I am trying to Migrate users data(including password and security questions) from 10g to 11gR2 what approach i have followed is..
    From 10g using API i retrieved users data including password and security questions and i stored all information into hashmap. This is one java program.
    And then i am trying to create that user in 11gR2 using API which i retrieved from 10g . From this 11g program i am creating object of 10g and i am using that hash map to retrieve user information.But i am not getting connection to 10g , it is throwing exception like unknown application server.Both sides i used API only as it is recommended to use API instead of JDBC connection.
    Help me in this regard ASAP and suggest if there is any other approach to Migrate users data.
    Thanks in Advance

    By using Trusted Recon, you won't be able to Fetch Password as it is.
    Since your goal is to fetch passwords too, please follow another approach.
    You won't be able to get connection to both 10g and 11g simultaneously in the same program.
    So, break this task in 2 phases. First connect with 10g, fetch user data in CSV format and then connect with 11G and read this CSV to create users.
    Once users are created properly, use APIs for creating challenge questions and answers.
    I think, you are getting exception like unknown application server because you are trying to connect to both 10g and 11g environments simultaneously.
    Follow the following steps:-
    (1) By using 10G APIsyou can't obtain password of user profile in decrypted form. So, Fetch password by using tcDataProvider. It will give you plain text password.
    (2) In a custom scheduler written in 10g, retrieve this data in CSV. After all you can't store this info in
    String query = "SELECT USR_LOGIN, USR_PASSWORD, USR_FIRST_NAME, USR_LAST_NAME FROM USR";//Add all fields which you want to retrieve from your 10G
    (3) Use this query, tcDataProvider, tcDataSet and Java I/O (or any other CSV Third Party tool like the ones obtained in csv.jar in XL_HOME/ext folder) fetch this info in a CSV.
    (4) Once CSV is generated, 10g machine is no more needed. Connect with 11g using 11g APIs. Write your custom 11G scheduler in order to read this CSV and use 11g APIs and create users for each record.
    (5) Once user records are created in 11g, the difficult part is done. Transfer the Security questions too by using this CSV technique.
    Please share results with us.

  • Moving data from old power pc with OS 8.6 to computer with OS X

    How can you move data from an old power pc to a new MAC? And then, how do you permanently erase the old hard disk before donating the computer?

    Which model PowerPC with 8.6 are we talking about? Does it have an Ethernet port? Do you have Broadband Ethernet? Do you have your own Router?
    Which new Mac are you moving to, and how does the iMac in your equipment line fit in to all of this?

  • Moving Data from old hard drive to new hard drive

    I'm posting this question in this category , I hope it's the right place.  Here's my situation and general information. I have a Toshiba L35 S2161 laptop, it's about 5 yrs old, running WinXP Service Pack 3.   just got the blue screen, wouldn't boot, tried everything and I only have the factory provided "recovery" disc that will wipe my drive and lose all my data if I use it.  I hadn't been doing regular backups, so I want to save as much if not all of my data as I can.  I don't want to spend time here going over everything I've tried (boot disc like Knoppix, etc) which didn't work, just need to get to the bottom line and make sure I'm doing the next steps correctly.
    Here's my plan:
    l.  Remove old hard drive from laptop
    2.  purchase new hard drive  - I've been told any drive will work, it doesn't have to be designed for a Toshiba.  I've been looking at best prices, etc. mainly Western Digital.  I may purchase a larger drive to at least get more memory , although I really didn't have that much on my old drive, but had a lot of apparent memory-related problems, freezes, etc.
    3.  Install new blank drive into laptop, using the Toshiba factory recovery disc to reformat the disc and get it ready to receive the data.
    4.  Insert the recovery disc into the drive, turn on the computer and HOPE that it boots it up and installs a fresh install on the new drive.  Not sure what the screen willl show at this point - I've never installed a new drive in a laptop, so will be somewhat nervous at this point to see what happens.
    5.  Assuming the new drive is recognized by the recovery disc and installs a clean version of WIN XP (I'm assuming that's what's on the recovery disc that came with my computer, because it had WIN XP pre-installed), I will now be ready to transfer data from the old drive.
    6.  This is the step I'm not sure what or how to do next.  I would like to clone my entire old drive onto the new drive but I'm afraid if I do that it might load damaged files that would mess up my new install.  I purchased a disc enclosure with a USB cable (haven't received it yet), and was going to use that somehow.  I presume I will plug it in and it will show up as an external drive and I can access that drive and transfer my priority files or data first and then try to transfer as much more as I can....
    I need further help and information about this last step...the critical data transfer method.  There are cloning kits that are plug and play and would transfer my entire system as it was before the blue screen, but again, I'm concerned about transferring everything onto a clean install and finding myself with a computer that won't boot again because it transferred a bad file.  Also, would it overwrite or interfere with the clean install on the new drive?   I'm stuck at this point....hopefully someone out there can walk me through this final step with the best option for transfer.  I'm trying to avoid purchasing pricey software like Acronis (which I looked at online and it seemed difficult to use, to figure out what to transfer). 
    My laptop is obviously out of warranty, being five years old, and I am not interested in taking it in to repair tech, I want to do this myself and I think I can with a few helpful tips.  Thanks for any help asap...

    Hi!
    The error message I received on the blue screen was "unmountable boot volume" which leads me to believe i have a corrupted MBR.  This fiasco happened when I turned off my computer when it was running a program (I know, bad thing to do).
    I'm waiting on a new hard drive with a 120 GB (my old drive was 80GB).  I was told not to go above 120 or I'd have partition problems. 
    The SATA drive enclosure powers up the old drive (I can hear the disc spinning) so the disc isn't dead.  If I''d have the full retail vesion of WinXP I would of course have used the recovery console, but I only had the toshiba recovery disc that came with my system.  Also, I couldn't burn a CD boot disc...didn't have access to another computer nor would a tech friend do it for me as he apparently made a disc for someone and their system crashed so I had no choice but to buy a new drive and try to transfer the data via the SATA external drive/usb cable.
    My big question now is:  Will the Toshiba recovery disc actually reformat my new drive?  Someone on another forum with a Toshiba laptop had purchased a new drive, used the recovery disc and the cd just went into a constant loop and never went into a full install.  I'm thinking that maybe on a new drive there isn't enough knowledge base there for the recovery disc to install and reformat a new disc.  That would be a bummer because then I'd have no choice but to purchase a retail version of WinXP which I was hoping not to have to do.  Thats' getting into some $$ when added to all the other thing I've tried to get this laptop up and running .  (A Knoppix live CD boot disc didn't work either).
    I don't know what service pack was on the old disc...possibly SP2, but if it was only SP1, I fear it's not going to have the ability to format a new disc.  The disc does have drivers, etc. but the question is, are they able to install the OS.  I was running XP3 when it crashed, so I know i would have to download that again.  I just want to get the darned system back up and running with my new drive and restore disc.  IF I can do that, then I'll try to transfer the data from the old drive via the external drive enclosure.
    I also decided against the clone....I was afraid I'd be cloning a bad disc or errors onto a new drive, and didnt' want to stress a possibly wonky drive even more than it possibly is. 
    So, that's where I'm at now!  Thanks so much for any further insight into this....something tells me the recovery disc isn't going to format my new drive but fingers crossed...more thoughts/feedback appreciated!

  • Moving data from DSO to cube

    Experts,
    We are in a situation where we might have to move data from cube to DSO. We doing POC and finding that even if DSO is set to overwrite, the data is summing up for same billing document. I would like to know:
    1) is there any o

    Hi,
    As the data in the cube is already in additive mode, if you get the data to a DSO in overwrtite mode, you will get an incorrect data.
    So in order to get the correct data,, it is always best to keep the DSO id additive mode.
    e.g
    Request ID Customer Quantity
    20               1000          15
    21               1000          10
    in this situation you would want to report
    Customer Quantity
    1000          25
    an ODS with overwrite mode, without the request ID as a KEY would give
    Customer Quantity
    1000          10
    an ODS with overwrite mode, with the request ID as a KEY would give
    Customer Quantity
    1000          25
    an ODS with additive mode, with/without the request ID as a KEY would give
    Customer Quantity
    1000          25
    Hope this helps
    -Vikram

  • Moving data from a dead G5 with serial ATA drives to a Mac Pro

    I have a dead Power Mac dual 2.7 GHz G5. The red light on the edge of the mother board is on - and the machine does not start up. It has been dying slowly - the dual monitors turn off when I send an image to my printer.
    It is under Apple Care - and today I purchased a Mac Pro quad 3.0 to replace it.
    But - I need some files from the 2 drives in the dead dual 2.7 G5. They are 250GB Serial ATA; 7200 rpm drives.
    From what I read here I cannot simply put the drives into the new Mac Pro.
    So - I think that I need a firewire enclosure for the 250GB Serial ATA; 7200 rpm drives that will let me get out the data.
    What is a good enclosure to do this - or is there another way? I can't send in the G5 for repair until I rescue the data from the drives and erase them.
    Dual 2.7 G5, 2 intel iMacs, 1 MacBook, 1 Powerbook, 18 G4 Macs, 3 iPods   Mac OS X (10.4.9)   I teach Digital Photography clases - on Macs

    You can install them - you just won't be able to boot
    from them.
    Don't move programs over - you should re-install them
    instead.
    That would be great. I don't need to boot from them - I just need to rescue some files.
    Can I erase them from the Mac pro before I send the G5 off for service?

Maybe you are looking for