Migrating Loops

I just upgraded from my PowerBook G4 to Mac Pro, and I'm transferring all my songs to the new computer. I'm having a bit of trouble migrating the loops I created into the fresh logic install. I can only find some of them in the Loops file under Library>audio. The ones that I find and drop in the new loop browser only partially show up. Anyone done this before? I need an idiot-proof version of the process.

I've tried a bunch of things but still no dice. Anyone?

Similar Messages

  • TFS Integration Platform Error while Syncing - Caught and converted, The request was aborted: The request was canceled

    Hello,
    We are doing a 2-way version control sync between TFS 2013 servers. 
    TFS A - Project A is getting synched with TFS B - Project A.
    TFS A - Project A has got all the code and files, whereas TFS B - Project A is empty.
    Previously we have synced TFS 2013 projects across servers with out any error or issues. Today we came across an issue (check log at end of post) and hope Microsoft can
    provide some guidance on this. Following is the problem we are facing today:
    We have configured the TFS Integration platform tool for VC - 2 way sync and started the migration. Current
    status is that the discovery, analysis is complete and migration is half way when I started to get such errors
    (check log at end of post) .
    Note that the migration has not stopped...
    After doing some research and googling I understand that the reason for this error is that by default if
    an individual file transfer cannot complete 16MB in 5 minutes, the operation will fail. And the solution to this is to modify
    chunk size in the file of "TfsMigrationShell.exe.config". I have few questions regarding this...
    Question 1:Is
    above understanding and solution correct? I also notice that this error appears for every huge file (approx >30MB), is there any timeout setting at the target server? Is this something to be changed at the Tool level config or the Target IIS level
    config?
    Question 2: Suppose
    I do not change the config setting, then huge files are skipped due to this error. However, after the migration of the other files is finished, then, will the tool retry to sync the skipped (huge) files? Or has it skipped them forever?
    Question 3:
    Suppose I modify
    chunk size in the file of "TfsMigrationShell.exe.config", then, which of the following do I need to do (give me the best solution among the 3 below)
    a) Do I need
    to restart the migration (by deleting and recreating the project on the 2nd TFS Server and re-starting migration)? 
    b) Do
    I need to create a new configuration and restart migration?
    c) Do I need to restart the sync service?
    The concern here is that we have got 3 more jobs (TFS
    Integration tool configurations) running parallelly that
    is syncing Work Items & VC of other TFS projects and want to make sure that restarting the TFS Integration Job Service will not have any negative impact on the other jobs as well as no negative impact on the current migration job...
    Microsoft, Please advice. Thanks!
    Log:
    [1/21/2015 10:55:39 AM] AsyncUploadFileState.Completed -- caught and converted:
    [1/21/2015 10:55:39 AM] Microsoft.TeamFoundation.VersionControl.Client.VersionControlException: C:\TfsIPData\10\ications\MDM Folder\Binaries\v9.5\DMS Web Service Release 1.5.zip: The request was aborted: The request was canceled. ---> System.Net.WebException: The request was aborted: The request was canceled. ---> System.IO.IOException: Cannot close stream until all bytes are written.
    [1/21/2015 10:55:39 AM] at System.Net.ConnectStream.CloseInternal(Boolean internalCall, Boolean aborting)
    [1/21/2015 10:55:39 AM] --- End of inner exception stack trace ---
    [1/21/2015 10:55:39 AM] at System.Net.ConnectStream.CloseInternal(Boolean internalCall, Boolean aborting)
    [1/21/2015 10:55:39 AM] at System.Net.ConnectStream.System.Net.ICloseEx.CloseEx(CloseExState closeState)
    [1/21/2015 10:55:39 AM] at System.Net.ConnectStream.Dispose(Boolean disposing)
    [1/21/2015 10:55:39 AM] at System.IO.Stream.Close()
    [1/21/2015 10:55:39 AM] at System.IO.Stream.Dispose()
    [1/21/2015 10:55:39 AM] at Microsoft.TeamFoundation.VersionControl.Client.FileUploader.UploadChunk(FileStream fileContentStream, Int64 end)
    [1/21/2015 10:55:39 AM] --- End of inner exception stack trace ---
    [1/21/2015 10:55:39 AM] at Microsoft.TeamFoundation.VersionControl.Client.Workspace.EndUploadFile(IAsyncResult asyncResult, PendingChange& pendingChange, Byte[]& uploadedContentBaselineFileGuid, Int64& uncompressedLength)
    [1/21/2015 10:55:39 AM] at Microsoft.TeamFoundation.VersionControl.Client.Workspace.AsyncUploadFileState.Completed(IAsyncResult asyncResult)

    Hi Charles,
    I have been doing in depth research and testing of this and here are my findings:
    Replies to your points:
    1) It failed with UploadChunkSize value change. Tried 1MB, 500KB, even 100KB. Fails with same error! Is there any other setting we can change for the timeouts at the target TFS server also?
    2) The huge files are not skipped. The migration loops on the failed files and doesn't continue to process the next lot of files. So the migration is blocked on the huge files.
    3) OK. Thanks. 
    Please see below the finding on when migration succeeds and when it fails:-
    Scenario 1:
    a) Where Migration fails:
    TFS A is in India and TFS B is in UK. Both TFS are on different domain. Migration tool is installed on same machine as TFS A. There is a firewall that allows the access from TFS A machine to TFS B machine. The migration tool fails to work in this scenario
    when it encounters a file size approx > 8MB. I tried setting the UploadChunkSize value to 100KB. Yet it failed!
    b) Where Migration works:
    However, if the files are approx less than 3MB then the migration works without any errors.
    Scenario 2 where TFS migration tool works:
    TFS A is in India and TFS B is also in India (located closeby). Both TFS are on different domain. Migration tool is installed on same machine as TFS A. There is no firewall in this case. No changes were made to UploadChunkSize value and files with size
    of 10's and 100's of MBs got synced without any issue.
    Questions
    Question 1: Can we conclude now that there is some issue with the firewall or network in case of scenario 1? As mentioned above Scenario 1 works if file size is less then approx 3MB.
    Question 2: Is there any timeout setting on the Target TFS server which can be set so that it waits for the file without timing out?
    Thanks for any help...

  • AR Credit Concept of Aging

    Hi People,
    I have to calculate the aging for open items in Accounts Receivable. The aging has to be done based on the baseline date from BSID. Now it is to be done like this:
    Current.
    Aging 1: 1-30 days
    Aging 2: 31-60
    Aging 3: 61-90
    Aging 4: over 90 days
    Now I am not able to really understand how to go about it. If somebody is good in FI, can you suggest something or give me a code snippet or something which would give me an idea of how to do it? I would appreciate help on this very much. Its urgent too.
    Thanks & regards,
    AM

    Hi,
    1)you can display and execute RFDOPR<i>00</i> with SE38 or SA38
    2)try this code (summation of Kunnr/aging-no.)
    REPORT zforum25 .
    *data
    DATA : BEGIN OF aging OCCURS 4,
    no(1),
    from(2),
    to(5),
    END OF aging.
    DATA: itab TYPE bsid OCCURS 0 WITH HEADER LINE.
    DATA:BEGIN OF atab OCCURS 0,
    kunnr LIKE bsid-kunnr,
    no LIKE aging-no,
    dmbtr LIKE bsid-dmbtr,
    END OF atab.
    DATA difday TYPE sy-dbcnt.
    *sel-screen
    PARAMETERS currdate LIKE sy-datum DEFAULT sy-datum.
    PARAMETERS bukrs LIKE itab-bukrs.
    SELECT-OPTIONS: kunnr FOR itab-kunnr,
                    budat FOR itab-budat DEFAULT '20050401' TO '20050630'.
    START-OF-SELECTION.
    *abbreviation !
      APPEND '10130' TO aging.
      APPEND '23160' TO aging.
      APPEND '36190' TO aging.
      APPEND '49199999' TO aging.
    *select data
      SELECT        * FROM  bsid INTO TABLE itab
             WHERE  bukrs  = bukrs
             AND    kunnr  IN kunnr
             AND    budat  IN budat.
    *migration
      LOOP AT itab.
        CLEAR atab.
    *correct sign
        IF itab-shkzg = 'H'.
          itab-dmbtr = itab-dmbtr * -1.
        ENDIF.
        MOVE-CORRESPONDING itab TO atab.
    *assigning age of invoices
        difday = currdate - itab-bldat."bldat = invoice date
        LOOP AT aging WHERE from LE difday
                        AND to   GE difday.
        ENDLOOP.
        IF sy-subrc = 0.
          MOVE aging-no TO atab-no.
        ELSE.
    *error -> check customizing of aging / bldat of invoice
          MESSAGE s001(00) WITH 'cannot assign' .
        ENDIF.
        COLLECT atab.
      ENDLOOP.
    *print
      SORT atab.
      LOOP AT atab.
        WRITE: / atab-kunnr, atab-no, atab-dmbtr.
      ENDLOOP.
    regards Andreas

  • How do I get rid of the Microsoft Setup Assistant loop? I migrated my software/documents from another laptop so don't have the disk to reinstall. Please help! Can't open any Microsoft Office software, like Word, and stuck in a loop?

    I migrated my software/documents from another laptop so don't have the disk to reinstall. Please help! Can't open any Microsoft Office software, like Word (for 2008), and stuck in a loop?
    Whenever I select Word Microsoft Setup Assistant appears, asks for feedback, then after selecting okay (both on saying yes or no to feedback) goes on to a registration page. When I click on this it says I've already registered so I just click okay, and then move on to a update page. After this, if I click on Word, the process repeats itself.
    As I said, I don't have the disk to reinstall, and can't find the Office Settings to delete as many pages have suggested I should try. Safe Boot restarting also hasn't worked... Really stuck and need Word very soon for work.
    If you can help, that would be great, and feel free to ask any questions about the situation as I'm not an expert here.
    Cheers,
    Jack

    First, export your contact from iCloud.com and save them on your computer in a safe spot some where (like you desktop).  Use this to help you do this: http://support.apple.com/kb/PH3606
    Next, on both of your devices, go to Settings > iCloud and turn on contacts and select Merge. Then turn off contacts and select 'Delete form my [device]' when prompted.
    Now go back to iCloud.com and select a contact (yes they will all be messed up again) and select Command+A on a Mac or Control+A on a PC to select all of the contacts.  Tap the delete key on your keyboard (or right click /control click a contact and select delete).
    You iPhone, iPad and iCloud.com should not be empty for contacts.
    Go back to Settings > iCloud on both devices and turn on contacts again (you should not see merge this time).
    Next, go back to iCloud.com and import your contacts (those exported .vcards).  You can either drag and drop them into the empty contacts list in your web browser, or you can use the gear icon to import.
    You cleaned up contacts should import correctly into iCloud.com and sync to both of your devices.
    Good luck.

  • Quick Migrate Creates a Trigger with Infinite Loop

    Quick Migrate did a good job at nicely converting my MS SQL Database to Oracle. The only problem that I have is triggers. I have a table with the companies and when the table is just created and there are no rows in the table, on the very first insert it goes into a constant loop. There is no code I have altered after the migration has been completed.
    So, there is:
    table: RA_COMPANY
    sequence: SEQ_RA_COMPANY_ID
    trigger: TRG_RA_COMPANY_ID
    table has no rows, trigger has been compiled and sequence hasn't been accessed so on the first insert, the trigger should put 1 into the newly added row. Here is the migration code for the trigger:
    CREATE OR REPLACE TRIGGER TRG_RA_COMPANY_ID BEFORE INSERT OR UPDATE ON RA_COMPANY
    FOR EACH ROW
    DECLARE
    v_newVal NUMBER(12) := 0;
    v_incval NUMBER(12) := 0;
    BEGIN
      IF INSERTING AND :new.ID IS NULL THEN
        SELECT SEQ_RA_COMPANY_ID.NEXTVAL INTO v_newVal FROM DUAL;
        -- If this is the first time this table have been inserted into (sequence == 1)
        IF v_newVal = 1 THEN
          --get the max indentity value from the table
          SELECT max(ID) INTO v_newVal FROM RA_COMPANY;
          v_newVal := v_newVal + 1;
          --set the sequence to that value
          LOOP
               EXIT WHEN v_incval>=v_newVal;
               SELECT SEQ_RA_COMPANY_ID.nextval INTO v_incval FROM dual;
          END LOOP;
        END IF;
       -- assign the value from the sequence to emulate the identity column
       :new.ID := v_newVal;
      END IF;
    END;
    ALTER TRIGGER TRG_RA_COMPANY_ID ENABLE;
    /and on the first insert, the loop that is in the middle will be an infinite loop.
    I do need only the inserts and I removed the UPDATE part and the loop itself but I am just wondering why is migration process creating this loop.
    thanks

    Hi Tridy,
    I look into this for you.
    I can see that this line
    SELECT max(ID) INTO v_newVal FROM RA_COMPANY;
    Is going to cause an issue if there are no rows.
    I wonder if
    SELECT NVL(max(ID),0) INTO v_newVal FROM RA_COMPANY;
    Solve the issue?
    Ill do some tests latter today and get back to you.
    Regards,
    Dermot.

  • Migration Assistant loops on Dupliate Account Found

    When I try to migrate and it finds Duplicate accounts, it seems that whatever I answer, when I click next, it goes back to Step 4 Machine Settings, and when I click migrate on that it just goes back to the duplicate account found dialogue, and so on forever.
    How should I proceed?

    OK, so what did happen when you created an account with the same name as an existing account.
    What user did you log in as? (the one with the same name or another).
    Did you check all the check boxes when asked what you wanted to migrate? In particular did you say you wanted to migrate Applications?
    When you clicked on migrate, did it bring up a dialogue box saying 'Duplicate Account Found'? Which accounts did it claim were duplicate? Which options were available? Was one of them greyed out? Which option did you select, and what then happened?
    Thanks, because I don't see how to migrate when I have duplicate accounts.

  • Migration Assistant Endless Loop?

    All my accounts on my machine including the adminstrator account launch automatically launch the migration assistant and upon completion of the assistant it logs you out. When you login again the migration assistant comes back up again. I have been unable to get past it. Has anyone experienced the same thing or been able to work around? This problem starting for me when I picked up the login and keychain update from apple today.

    Yep. I can't figure out how to fix it, either.
    I can log into the machine remotely via SSH on the command line just fine. Everything is there. But the update doesn't seem to have left anything in the Receipts directory, so I can't figure out exactly what it changed on the system, to try to fix it.

  • For loop code migration

    I found this code in an AS2 .fla and would like to use it in AS3.  (The idea is to throw a random number of copies of the movieclip "star_mc" on the stage via AS). 
    for (var i =0;i<50;i++) {
    star_mc.duplicateMovieClip("star"+i,i,{_x:Math.random()*350,_y:Math.random()*150});
    this["star"+i].gotoAndPlay(Math.random());
    I get this error message:  TypeError: Error #1006: duplicateMovieClip is not a function.
    at many_fla::MainTimeline/frame1()

    The duplicateMovieClip method was removed in AS3 and does not have an equivalent directly replacing it.  In order to do what you want you will have to assign a class identifier to the object in the library (via the Linkage option) and use that to create new instances...
    for (var i:int=0;i<50;i++) {
         var star:Star = new Star();
         star.name = "star"+String(i);
         star.x = Math.random()*350;
         star.y:Math.random()*150);
         addChild(star);
         star.gotoAndPlay(1);

  • Item Conversion Template Issue for Oracle Migration - Copy functionality

    Hi,
    I am working on Migration project which is from Radius ERP to Oracle 11.5.10.2.
    Currently working on Item Conversion. This Item conversion having the fileds like (ORGANIZATION_ID,SEGMENT1,DESCRIPTION,ITEM_TYPE,COST_OF_SALES_ACCOUNT,SALES_ACCOUNT,ATTRIBUTE_CATEGORY,ATTRIBUTE1,ATTRIBUTE2,ATTRIBUTE3,ATTRIBUTE4,ATTRIBUTE5,ATTRIBUTE6,ATTRIBUTE7,ATTRIBUTE8,ATTRIBUTE9,ATTRIBUTE10,ATTRIBUTE11,ATTRIBUTE12,ATTRIBUTE13,ATTRIBUTE14,ATTRIBUTE15,GLOBAL_ATTRIBUTE10,REF_INVENTORY_ITEM_ID,REF_ORGANIZATION_ID).
    I have validated those fields and loaded into interfacing to Oracle Successfully in the master Org and Validated through Frontend.
    when I close the form, I will be receiving the below warning message and telling that Template id needs to be assign to the item before assigning item to the Org. The message is showing like
    "*You have not applied a template to this item, please apply a template before assigning this item to an ORG.*"
    Please find the below package which I wrote for this conversion requirement.
    The customer is saying like need to achieve the copy functionality based on the "REF_INVENTORY_ITEM_ID and REF_ORGANIZATION_ID".
    The Issue is am not able to handle the copy functionality and getting above message. kindly refer the package and suggest me where i am doing the mistake. Its high priority issue for me.
    Thanks in advance.
    CREATE OR REPLACE PACKAGE APPS.xxxx_inv_items_conv_pkg
    AS
    PROCEDURE xxx_item_conversion_proc (
    errbuf OUT VARCHAR2,
    retcode OUT VARCHAR2,
    p_org_id IN NUMBER,
    -- p_commit_point IN NUMBER,
    p_load_code IN VARCHAR2
    IS
    <<Local Variables Declaration>> <<For space limit deleted these variables>>
    CURSOR cur_item_master (pc_org_code VARCHAR2)
    IS
    SELECT itemstg.*
    FROM xxx_inv_system_items_stg itemstg
    WHERE itemstg.organization_id = pc_org_code
    AND itemstg.status_flag IS NULL;
    CURSOR cur_item_master_dup (pc_org_code VARCHAR2)
    IS
    SELECT itemstg.segment1, itemstg.organization_id
    FROM xxx_inv_system_items_stg itemstg
    WHERE itemstg.ROWID <
    (SELECT MAX (b.ROWID)
    FROM xxx_inv_system_items_stg b
    WHERE b.segment1 = itemstg.segment1
    AND b.organization_id = itemstg.organization_id
    AND b.organization_id = pc_org_code
    AND itemstg.status_flag IS NULL
    AND b.status_flag IS NULL);
    BEGIN
    IF p_load_code = 'Insert'
    THEN
    l_transaction_type := 'CREATE'; -- Default Value in I/F Table
    ELSIF p_load_code = 'Update'
    THEN
    l_transaction_type := 'UPDATE'; -- Default Value in I/F Table
    END IF;
    DBMS_OUTPUT.put_line ( 'Validation Starts At :'
    || TO_CHAR (SYSDATE, 'DD-MON-YYYY HH24:MI:SS')
    --Checking for Duplicate Records items
    BEGIN
    UPDATE xxx_inv_system_items_stg a
    SET a.status_flag = 'E',
    a.error_message = 'Duplicate Record'
    WHERE a.ROWID >
    ANY (SELECT b.ROWID
    FROM xxx_inv_system_items_stg b
    WHERE a.segment1 = b.segment1
    AND a.organization_id = b.organization_id)
    AND a.organization_id = p_org_id
    AND a.status_flag IS NULL;
    EXCEPTION
    WHEN OTHERS
    THEN
    DBMS_OUTPUT.put_line ('Exception in updating duplicates'
    || SQLERRM
    END;
    DBMS_OUTPUT.put_line ('CheckPoint: Duplicate Record');
    BEGIN
    SELECT organization_id
    INTO l_organization_id
    FROM org_organization_definitions
    WHERE organization_id = p_org_id;
    EXCEPTION
    WHEN NO_DATA_FOUND
    THEN
    DBMS_OUTPUT.PUT_LINE(p_org_id||' Org Does Not Exist');
    -- p_retcode := '2';
    WHEN OTHERS
    THEN
    DBMS_OUTPUT.PUT_LINE('Exception in Getting Org Id'||'Cannot Proceed');
    -- p_retcode := '2';
    END ;
    DBMS_OUTPUT.put_line ('CheckPoint: Orgcode' || l_organization_id);
    /* IF p_retcode = '2'
    THEN
    RETURN;
    END IF;*/
    --Block             : Setting Master and Validation Orgs Flags
    BEGIN
    SELECT DECODE (master_organization_id, l_organization_id, 'Y', 'N'),
    master_organization_id
    INTO l_master_org,
    l_master_org_id
    FROM mtl_parameters
    WHERE organization_id = l_organization_id;
    EXCEPTION
    WHEN OTHERS
    THEN
    l_master_org := 'N';
    l_master_org_id := NULL;
    DBMS_OUTPUT.PUT_LINE(p_org_id||' Org Does Not Exist');
    END ;
    DBMS_OUTPUT.put_line ('master_organization_id');
    SELECT fnd_profile.VALUE ('USER_ID')
    INTO l_user_id
    FROM DUAL;
    -- Block : Set the SET_PROCESS_ID
    l_set_process_id := l_organization_id;
    l_insert_count := 0;
    LOOP
    BEGIN
    SELECT COUNT (segment1)
    INTO l_insert_count
    FROM mtl_system_items_interface
    WHERE set_process_id = l_set_process_id
    AND transaction_type = l_transaction_type
    AND process_flag = 1;
    EXCEPTION
    WHEN OTHERS
    THEN
    l_insert_count := 0;
    END;
    EXIT WHEN l_insert_count = 0;
    l_set_process_id := l_set_process_id + 10;
    END LOOP;
    DBMS_OUTPUT.put_line ('SET PROCESS ID -l_insert_count ' || l_insert_count);
    IF p_load_code = 'Insert'
    THEN
    BEGIN
    FOR recitem_data IN cur_item_master_dup (p_org_id)
    LOOP
    UPDATE xxx_inv_system_items_stg
    SET status_flag = l_processed_flag,
    error_message = l_error_message
    WHERE segment1 = recitem_data.segment1
    AND organization_id = recitem_data.organization_id
    AND status_flag IS NULL;
    COMMIT;
    END LOOP;
    EXCEPTION
    WHEN OTHERS
    THEN
    DBMS_OUTPUT.put_line ( 'Update of Duplicates Failed : '
    || SQLCODE
    || '--'
    || SQLERRM
    END ;
    END IF;
    FOR recitem_data IN cur_item_master (p_org_id)
    LOOP
    <<Local Variables Declaration>> <<For space limit deleted these variables>>
    IF p_load_code = 'Update'
    THEN
    l_error_message := 'Update Mode' || l_error_delimiter;
    END IF;
    --l_count := -1;
    BEGIN
    DBMS_OUTPUT.put_line ('CheckPoint: ItemValidationStart');
    SELECT DISTINCT inventory_item_id,
    restrict_subinventories_code,
    restrict_locators_code
    INTO l_inventory_item_id,
    l_restrict_subinventories_code,
    l_restrict_locators_code
    FROM apps.mtl_system_items_b msi
    WHERE msi.organization_id = l_organization_id
    AND msi.segment1 = UPPER (recitem_data.segment1);
    EXCEPTION
    WHEN NO_DATA_FOUND
    THEN
    l_count := 0;
    -- l_processed_flag := 'F';
    DBMS_OUTPUT.put_line ('inventory_item_id - AFTER MAIN LOOP' || l_inventory_item_id||'-'||l_processed_flag);
    DBMS_OUTPUT.put_line ('inventory_item_id - AFTER MAIN LOOP' || l_inventory_item_id||'-'||recitem_data.ref_inventory_item_id);
    WHEN OTHERS
    THEN
    l_count := -1;
    -- l_processed_flag := 'F';
    DBMS_OUTPUT.put_line ('inventory_item_id - AFTER MAIN LOOP' || l_inventory_item_id||'-'||l_processed_flag);
    END ;
    DBMS_OUTPUT.put_line ('inventory_item_id - AFTER MAIN LOOP'||recitem_data.ref_inventory_item_id);
    IF l_count = -1
    THEN
    l_processed_flag := 'F';
    l_error_message :=
    l_error_message
    || 'Exception - Checking Item already Present'
    || l_error_delimiter;
    DBMS_OUTPUT.put_line ('Exception - Checking Item already Present'||l_processed_flag);
    ELSIF (l_count > 0 AND p_load_code = 'Insert')
    THEN
    l_processed_flag := 'F';
    l_error_message :=
    l_error_message
    || 'Item Already Exists In '
    || p_org_id
    || ' Organization '
    || l_error_delimiter;
    DBMS_OUTPUT.put_line ('Exception - Checking Item already Present1'||l_processed_flag);
    ELSIF (l_count = 0 AND p_load_code = 'Update')
    THEN
    l_processed_flag := 'F';
    l_error_message :=
    l_error_message
    || 'Item Not Present In '
    || p_org_id
    || ' Organization '
    || l_error_delimiter;
    DBMS_OUTPUT.put_line ('Exception - Checking Item already Present2'||l_processed_flag);
    END IF;
    BEGIN
    SELECT count(*)
    INTO l_seg_count
    FROM apps.mtl_system_items_b msi
    WHERE msi.organization_id = l_organization_id
    AND msi.segment1 = UPPER (recitem_data.segment1);
    EXCEPTION
    WHEN NO_DATA_FOUND
    THEN
    l_seg_count := 0;
    l_processed_flag:='F';
    l_error_message :=
    l_error_message
    || 'Item Not Present In Oracle'
    || p_org_id
    || ' Organization '
    || l_error_delimiter;
    DBMS_OUTPUT.put_line ('inventory_item_id - l_seg_count ' ||l_seg_count||'-'||l_processed_flag );
    WHEN OTHERS
    THEN
    l_seg_count := -1;
    l_processed_flag:='F';
    l_error_message :=
    l_error_message
    || 'Item Not Present In Oracle'
    || p_org_id
    || ' Organization '
    || l_error_delimiter;
    END ;
    IF p_load_code = 'Insert'
    THEN
    IF l_organization_id != l_master_org_id
    THEN
    BEGIN
    SELECT COUNT (1)
    INTO l_org_item_count
    FROM apps.mtl_system_items_b msi
    WHERE msi.organization_id = l_master_org_id
    AND msi.segment1 = UPPER (recitem_data.segment1);
    EXCEPTION
    WHEN NO_DATA_FOUND
    THEN
    l_org_item_count := 0;
    DBMS_OUTPUT.put_line ('l_org_item_count ' ||l_org_item_count||'-'||l_processed_flag );
    WHEN OTHERS
    THEN
    l_org_item_count := -1;
    END ;
    IF l_org_item_count = -1
    THEN
    l_processed_flag := 'F';
    l_error_message :=
    l_error_message
    || 'Exception - Checking Item in Master '
    || l_error_delimiter;
    ELSIF l_org_item_count = 0
    THEN
    l_processed_flag := 'F';
    l_error_message :=
    l_error_message
    || 'Item Does Not Exist in Master '
    || l_error_delimiter;
    END IF;
    END IF;
    END IF;
    DBMS_OUTPUT.put_line ( 'CheckPoint: Iteminmasterorg'
    || l_org_item_count
    IF (recitem_data.description IS NULL AND p_load_code = 'Insert')
    THEN
    l_processed_flag := 'F';
    l_error_message :=
    l_error_message || 'Description is NULL' || l_error_delimiter;
    END IF;
    IF (recitem_data.sales_account IS NOT NULL)
    THEN
    BEGIN
    SELECT code_combination_id
    INTO l_sales_account
    FROM gl_code_combinations_kfv
    WHERE code_combination_id= recitem_data.sales_account;
    EXCEPTION
    WHEN NO_DATA_FOUND
    THEN
    -- l_processed_flag := 'F'; --QUESTION
    l_error_message :=
    l_error_message
    || 'Sales Account Not Setup '
    || l_error_delimiter;
    DBMS_OUTPUT.put_line ( 'CheckPoint: Salesacct_Validation'
    || l_sales_account||'-'||l_processed_flag
    WHEN OTHERS
    THEN
    l_processed_flag := 'F';
    l_error_message :=
    l_error_message
    || 'Sales Account Exception '
    || SQLERRM
    || l_error_delimiter;
    DBMS_OUTPUT.put_line ( 'CheckPoint: Salesacct_Validation'
    || l_sales_account||'-'||l_processed_flag
    END ;
    END IF;
    IF (recitem_data.cost_of_sales_account IS NOT NULL)
    THEN
    BEGIN
    SELECT code_combination_id
    INTO l_cost_of_sales_account
    FROM gl_code_combinations_kfv
    WHERE code_combination_id =
    recitem_data.cost_of_sales_account;
    EXCEPTION
    WHEN NO_DATA_FOUND
    THEN
    l_processed_flag := 'F';
    l_error_message :=
    l_error_message
    || 'COGS Account Not Setup '
    || l_error_delimiter;
    DBMS_OUTPUT.put_line ( 'l_cost_of_sales_account'
    || l_cost_of_sales_account||'-'||l_processed_flag
    WHEN OTHERS
    THEN
    l_processed_flag := 'F';
    l_error_message :=
    l_error_message
    || 'COGS Account Exception '
    || SQLERRM
    || l_error_delimiter;
    DBMS_OUTPUT.put_line ( 'l_cost_of_sales_account'
    || l_cost_of_sales_account||'-'||l_processed_flag
    END check_cogs_account;
    END IF;
    BEGIN
    select distinct organization_code
    into l_ref_org_code
    from org_organization_definitions
    where organization_id = recitem_data.ref_organization_id ;
    EXCEPTION
    WHEN OTHERS THEN
    l_ref_org_code:= NULL;
    l_processed_flag:='F';
    l_error_message :=
    l_error_message
    || 'Reference Org Not Present'
    || p_org_id
    || ' Organization '
    || l_error_delimiter ;
    DBMS_OUTPUT.put_line ('l_ref_org_code'|| l_ref_org_code||'-'||l_processed_flag);
    END;
    BEGIN
    select concatenated_segments
    into l_ref_inventory_item_code
    from mtl_system_items_kfv
    where inventory_item_id = recitem_data.ref_inventory_item_id
    and organization_id = recitem_data.ref_organization_id ;
    EXCEPTION
    WHEN OTHERS THEN
    l_ref_inventory_item_code :=NULL;
    l_processed_flag:='F';
    l_error_message :=
    l_error_message
    || 'Reference Item Not Present'
    || p_org_id
    || ' Organization '
    || l_error_delimiter;
    DBMS_OUTPUT.put_line ('l_ref_inventory_item_code'|| l_ref_inventory_item_code||'-'||l_processed_flag);
    END;
    IF (l_master_org = 'N' AND p_load_code = 'Insert')
    THEN
    BEGIN
    SELECT count(*)
    INTO l_description_count
    FROM mtl_system_items_tl
    WHERE organization_id = l_master_org_id
    AND inventory_item_id =
    (SELECT inventory_item_id
    FROM mtl_system_items_b
    WHERE organization_id = l_master_org_id
    AND segment1 = recitem_data.segment1);
    END ;
    IF l_description_count > 0 THEN
    l_processed_flag:='F';
    l_error_message :=
    l_error_message
    || 'Item Description Not Present'
    || p_org_id
    || ' Organization '
    || l_error_delimiter;
    END IF;
    END IF;
    IF l_processed_flag = 'S'
    THEN
    BEGIN
    INSERT INTO mtl_system_items_interface
    (organization_id,
    segment1,
    description,
    ITEM_TYPE,
    COST_OF_SALES_ACCOUNT,
    SALES_ACCOUNT,
    attribute_category,
    set_process_id,
    transaction_type,
    process_flag,
    copy_organization_code,
    copy_item_number,
    creation_date,
    created_by,
    last_updated_by,
    last_update_date
    --attribute_category
    , attribute1
    , attribute2
    ,attribute3
    ,attribute4
    ,attribute5
    ,attribute6
    ,attribute7
    ,attribute8
    ,attribute9
    ,attribute10
    ,attribute11
    ,attribute12
    ,attribute13
    ,attribute14
    ,attribute15
    ,global_attribute10
    VALUES (l_organization_id,
    recitem_data.segment1,
    recitem_data.description,
    recitem_data.ITEM_TYPE,
    recitem_data.COST_OF_SALES_ACCOUNT,
    recitem_data.SALES_ACCOUNT,
    recitem_data.ATTRIBUTE_CATEGORY,
    l_set_process_id,
    l_transaction_type, --,l_transaction_type
    l_process_flag,
    l_ref_org_code,
    l_ref_inventory_item_code,
    SYSDATE, l_user_id,
    l_user_id, SYSDATE
    -- l_attribute_category
    ,recitem_data.attribute1
    ,recitem_data.attribute2
    ,recitem_data.attribute3
    ,recitem_data.attribute4
    ,recitem_data.attribute5
    ,recitem_data.attribute6
    ,recitem_data.attribute7
    , recitem_data.attribute8
    ,recitem_data.attribute9
    , recitem_data.attribute10
    ,recitem_data.attribute11
    ,recitem_data.attribute12
    ,recitem_data.attribute13
    ,recitem_data.attribute14
    ,recitem_data.attribute15
    , substr(recitem_data.global_attribute10,1,length(recitem_data.global_attribute10)-1) --recitem_data.global_attribute10
    l_insert_count := l_insert_count + 1;
    /* IF (l_insert_count = NVL (p_commit_point, 10000))
    THEN
    -- l_set_process_id := l_set_process_id + 10; -- REVERT BACK CHANGE
    l_insert_count := 0;
    END IF;
    EXCEPTION
    WHEN OTHERS
    THEN
    l_processed_flag := 'F';
    l_error_message :=
    l_error_message
    || 'Error in Inserting Item '
    || SQLERRM
    || l_error_delimiter;
    END ;
    COMMIT;
    DBMS_OUTPUT.put_line ('CheckPoint: Insertion Completed in Intfacetable');
    END IF;
    BEGIN
    UPDATE xxx_inv_system_items_stg
    SET status_flag = l_processed_flag,
    error_message = l_error_message
    WHERE segment1 = recitem_data.segment1
    AND organization_id = recitem_data.organization_id
    AND status_flag IS NULL;
    COMMIT;
    DBMS_OUTPUT.put_line ('Error Message'||l_error_message);
    EXCEPTION
    WHEN OTHERS
    THEN
    fnd_file.put_line (fnd_file.LOG,
    'Error:Updating Item:'
    || SQLCODE
    || '-'
    || SQLERRM
    END;
    END LOOP; --FOR recitem_data IN cur_item_master
    DBMS_OUTPUT.put_line ( 'Validation Ends At :'
    || TO_CHAR (SYSDATE, 'DD-MON-YYYY HH24:MI:SS')
    DBMS_OUTPUT.put_line ('Number of records inserted into Table Successfully -->'|| l_insert_count);
    END;
    END xxxx_inv_items_conv_pkg;
    Edited by: 896170 on Apr 12, 2013 11:58 PM

    Issue got solved... changed the PO line amount as:
    <?xdoxslt:set_variable($_XDOCTX,'line_amt',xdoxslt:to_number(LINE_AMOUNT))?>
    Reference :Syntax for 'to_number'
    Regards
    Manikanta Panigrahi

  • Elapsed time went up from 1min to 22min after migrating from 10g to 11g

    I just migrated one of my database from 10.2.0.2(Red hat Linux, 2 node RAC, sga= 1Gb) to 11.2.0.1 (red Hat Linux 2 Node RAC, SGA=7GB)
    The timing for one of the specific query shoot up from 1min to 22 min.
    Following is the query:
    SELECT /*+ gather_plan_statistics */   docr.DRCONTENT
        FROM      WRPADMIN.T_DOCREPORT docr, WRPADMIN.t_document doc
               WHERE doc.docid = docr.docid
       AND 294325 = doc.rdocid
               AND ( ( ( (EXISTS
                             (SELECT   'X'
                                FROM   WRPADMIN.t_mastermap mstm1,
                                       WRPADMIN.t_docdimmap docdim1
                               WHERE       doc.docid = mstm1.docid
                                       AND mstm1.dimlvlid = 2
                                       AND mstm1.mstmapid = docdim1.mstmapid
                                       AND docdim1.dimid IN (86541))))
                      OR (EXISTS
                             (SELECT   'X'
                                FROM   WRPADMIN.t_mastermap mstm2,
                                       WRPADMIN.t_docdimmap docdim2
                               WHERE       doc.rdocid = mstm2.rdocid
                                       AND mstm2.dimlvlid = 1
                                       AND mstm2.mstmapid = docdim2.mstmapid
                                       AND docdim2.dimid IN (28388)))))
    ORDER BY   doc.DOCIDThe select field (docr.DRCONTENT) is a CLOB column.
    Following is the plan and statistics in 10g
    Statistics
              1  recursive calls
              0  db block gets
         675018  consistent gets
          52225  physical reads
              0  redo size
       59486837  bytes sent via SQL*Net to client
       27199426  bytes received via SQL*Net from client
         103648  SQL*Net roundtrips to/from client
              1  sorts (memory)
              0  sorts (disk)
          51823  rows processed
    SQL>
    Plan hash value: 129748299
    | Id  | Operation                               | Name           | Starts | E-Rows | A-Rows |   A-Time   | Buffers | Reads  |  OMem |  1Mem | Used-Mem |
    |   1 |  SORT ORDER BY                          |                |      1 |     50 |  51823 |00:00:14.72 |     627K|   5379 |    26M|  1873K|   23M (0)|
    |*  2 |   FILTER                                |                |      1 |        |  51823 |00:00:08.90 |     627K|   5379 |       |       |          |
    |   3 |    TABLE ACCESS BY GLOBAL INDEX ROWID   | T_DOCREPORT    |      1 |      1 |  51823 |00:00:05.42 |     159K|   3773 |       |       |          |
    |   4 |     NESTED LOOPS                        |                |      1 |     50 |    103K|00:00:12.65 |     156K|    628 |       |       |          |
    |   5 |      TABLE ACCESS BY GLOBAL INDEX ROWID | T_DOCUMENT     |      1 |     50 |  51823 |00:00:00.15 |     481 |    481 |       |       |          |
    |*  6 |       INDEX RANGE SCAN                  | RDOC2_INDEX    |      1 |    514 |  51823 |00:00:00.09 |     245 |    245 |       |       |          |
    |*  7 |      INDEX RANGE SCAN                   | DOCID9_INDEX   |  51823 |      1 |  51823 |00:00:00.46 |     155K|    147 |       |       |          |
    |*  8 |    TABLE ACCESS BY GLOBAL INDEX ROWID   | T_DOCDIMMAP    |  51823 |      1 |      0 |00:00:04.52 |     467K|   1140 |       |       |          |
    |   9 |     NESTED LOOPS                        |                |  51823 |      1 |    207K|00:00:03.48 |     415K|    479 |       |       |          |
    |* 10 |      TABLE ACCESS BY GLOBAL INDEX ROWID | T_MASTERMAP    |  51823 |      1 |  51823 |00:00:01.20 |     207K|    190 |       |       |          |
    |* 11 |       INDEX RANGE SCAN                  | DOCID4_INDEX   |  51823 |      1 |  51824 |00:00:00.41 |     155K|    146 |       |       |          |
    |* 12 |      INDEX RANGE SCAN                   | MSTMAPID_INDEX |  51823 |      1 |    103K|00:00:00.43 |     207K|    289 |       |       |          |
    |* 13 |     TABLE ACCESS BY GLOBAL INDEX ROWID  | T_DOCDIMMAP    |      1 |      1 |      1 |00:00:01.05 |     469 |    466 |       |       |          |
    |  14 |      NESTED LOOPS                       |                |      1 |      1 |     15 |00:00:14.62 |     468 |    465 |       |       |          |
    |* 15 |       TABLE ACCESS BY GLOBAL INDEX ROWID| T_MASTERMAP    |      1 |      1 |      1 |00:00:01.02 |     464 |    463 |       |       |          |
    |* 16 |        INDEX RANGE SCAN                 | RDOCID3_INDEX  |      1 |    629 |  44585 |00:00:00.29 |     198 |    198 |       |       |          |
    |* 17 |       INDEX RANGE SCAN                  | MSTMAPID_INDEX |      1 |      1 |     14 |00:00:00.02 |       4 |      2 |       |       |          |
    Predicate Information (identified by operation id):
       2 - filter(( IS NOT NULL OR  IS NOT NULL))
       6 - access("DOC"."RDOCID"=294325)
       7 - access("DOC"."DOCID"="DOCR"."DOCID")
       8 - filter("DOCDIM1"."DIMID"=86541)
      10 - filter("MSTM1"."DIMLVLID"=2)
      11 - access("MSTM1"."DOCID"=:B1)
      12 - access("MSTM1"."MSTMAPID"="DOCDIM1"."MSTMAPID")
      13 - filter("DOCDIM2"."DIMID"=28388)
      15 - filter("MSTM2"."DIMLVLID"=1)
      16 - access("MSTM2"."RDOCID"=:B1)
      17 - access("MSTM2"."MSTMAPID"="DOCDIM2"."MSTMAPID")following is the plan in 11g:
    Statistics
             32  recursive calls
              0  db block gets
       20959179  consistent gets
         105948  physical reads
            348  redo size
       37320945  bytes sent via SQL*Net to client
       15110877  bytes received via SQL*Net from client
         103648  SQL*Net roundtrips to/from client
              3  sorts (memory)
              0  sorts (disk)
          51823  rows processed
    SQL>
    Plan hash value: 1013746825
    | Id  | Operation                               | Name           | Starts | E-Rows | A-Rows |   A-Time   | Buffers | Reads  |  OMem |  1Mem | Used-Mem |
    |   0 | SELECT STATEMENT                        |                |      1 |        |  51823 |00:01:10.08 |      20M|   2306 |       |       |          |
    |   1 |  SORT ORDER BY                          |                |      1 |      1 |  51823 |00:01:10.08 |      20M|   2306 |  9266K|  1184K| 8236K (0)|
    |*  2 |   FILTER                                |                |      1 |        |  51823 |00:21:41.79 |      20M|   2306 |       |       |          |
    |   3 |    NESTED LOOPS                         |                |      1 |        |  51823 |00:00:01.95 |    8054 |   1156 |       |       |          |
    |   4 |     NESTED LOOPS                        |                |      1 |    335 |  51823 |00:00:00.99 |    4970 |    563 |       |       |          |
    |   5 |      TABLE ACCESS BY GLOBAL INDEX ROWID | T_DOCUMENT     |      1 |    335 |  51823 |00:00:00.38 |     402 |    401 |       |       |          |
    |*  6 |       INDEX RANGE SCAN                  | RDOC2_INDEX    |      1 |    335 |  51823 |00:00:00.17 |     148 |    147 |       |       |          |
    |*  7 |      INDEX RANGE SCAN                   | DOCID9_INDEX   |  51823 |      1 |  51823 |00:00:00.55 |    4568 |    162 |       |       |          |
    |   8 |     TABLE ACCESS BY GLOBAL INDEX ROWID  | T_DOCREPORT    |  51823 |      1 |  51823 |00:00:00.94 |    3084 |    593 |       |       |          |
    |   9 |    CONCATENATION                        |                |  51823 |        |  51823 |00:22:16.08 |      20M|   1150 |       |       |          |
    |  10 |     NESTED LOOPS                        |                |  51823 |        |      0 |00:00:02.71 |     221K|   1150 |       |       |          |
    |  11 |      NESTED LOOPS                       |                |  51823 |      1 |    103K|00:00:01.19 |     169K|    480 |       |       |          |
    |* 12 |       TABLE ACCESS BY GLOBAL INDEX ROWID| T_MASTERMAP    |  51823 |      1 |  51823 |00:00:00.72 |     108K|    163 |       |       |          |
    |* 13 |        INDEX RANGE SCAN                 | DOCID4_INDEX   |  51823 |      1 |  51824 |00:00:00.52 |   56402 |    163 |       |       |          |
    |* 14 |       INDEX RANGE SCAN                  | MSTMAPID_INDEX |  51823 |      2 |    103K|00:00:00.60 |   61061 |    317 |       |       |          |
    |* 15 |      TABLE ACCESS BY GLOBAL INDEX ROWID | T_DOCDIMMAP    |    103K|      1 |      0 |00:00:01.14 |   52584 |    670 |       |       |          |
    |  16 |     NESTED LOOPS                        |                |  51823 |        |  51823 |00:22:13.19 |      20M|      0 |       |       |          |
    |  17 |      NESTED LOOPS                       |                |  51823 |      1 |    725K|00:22:12.31 |      20M|      0 |       |       |          |
    |* 18 |       TABLE ACCESS BY GLOBAL INDEX ROWID| T_MASTERMAP    |  51823 |      1 |  51823 |00:22:11.09 |      20M|      0 |       |       |          |
    |* 19 |        INDEX RANGE SCAN                 | RDOCID3_INDEX  |  51823 |    336 |   2310M|00:12:08.04 |    6477K|      0 |       |       |          |
    |* 20 |       INDEX RANGE SCAN                  | MSTMAPID_INDEX |  51823 |      2 |    725K|00:00:00.83 |   51838 |      0 |       |       |          |
    |* 21 |      TABLE ACCESS BY GLOBAL INDEX ROWID | T_DOCDIMMAP    |    725K|      1 |  51823 |00:00:00.92 |   51823 |      0 |       |       |          |
    Predicate Information (identified by operation id):
       2 - filter( IS NOT NULL)
       6 - access("DOC"."RDOCID"=294325)
       7 - access("DOC"."DOCID"="DOCR"."DOCID")
      12 - filter("MSTM1"."DIMLVLID"=2)
      13 - access("MSTM1"."DOCID"=:B1)
      14 - access("MSTM1"."MSTMAPID"="DOCDIM1"."MSTMAPID")
      15 - filter((INTERNAL_FUNCTION("DOCDIM1"."DIMID") AND (("DOCDIM1"."DIMID"=86541 AND "MSTM1"."DIMLVLID"=2 AND "MSTM1"."DOCID"=:B1) OR
                  ("DOCDIM1"."DIMID"=28388 AND "MSTM1"."DIMLVLID"=1 AND "MSTM1"."RDOCID"=:B2))))
      18 - filter(("MSTM1"."DIMLVLID"=1 AND (LNNVL("MSTM1"."DOCID"=:B1) OR LNNVL("MSTM1"."DIMLVLID"=2))))
      19 - access("MSTM1"."RDOCID"=:B1)
      20 - access("MSTM1"."MSTMAPID"="DOCDIM1"."MSTMAPID")
      21 - filter((INTERNAL_FUNCTION("DOCDIM1"."DIMID") AND (("DOCDIM1"."DIMID"=86541 AND "MSTM1"."DIMLVLID"=2 AND "MSTM1"."DOCID"=:B1) OR
                  ("DOCDIM1"."DIMID"=28388 AND "MSTM1"."DIMLVLID"=1 AND "MSTM1"."RDOCID"=:B2))))Calling all performance experts. Any ideas ??
    Edited by: dm_ptldba on Oct 8, 2012 7:50 AM

    If you check lines 2, 3, 8, and 13 in the 10g plan you will see that Oracle has operated your two EXISTS subqueries separately (there is a bug with multiple filter subqueries in that version that indents each subquery after the first one extra place, so the shape of the plan is a little deceptive). The statistics show that the second subquery only ran once because existence was almost always satistfied by the first.
    In the 11g plan, lines 2, 3, and 9 show that the optimizer has transformed your TWO subqueries into a single subquery, then turned transformed the single subquery into a concatenation and this has, in effect, made it execute both subqueries for every row from the driving table - all the extra work appears from the redundant execution of the thing that was the second EXISTS subquery.
    If you extract the OUTLINE from the execution plans (add 'outline' to the call to dbms_xplan as one of the format options) you may see some hint that shows the optimizer combining the two subqueries - if so then put in the "NO_xxx" hint to block it. Alternatively you could simply try adding the hint stop ALL cost-based query transformations /*+ no_query_transformation */
    Regards
    Jonathan Lewis

  • Query is taking long time to execute after migrating to 10g r2

    Hi
    We recently migrated the database from 9i to 10gr2 ((10.2.0.2.0).. This query was running in acceptable time before the upgrade in 9i.. Now it is taking a long long time to execute this... Can you please let me know what should i do to improve the performance now.. We are running stats everyday..
    Thanks for your help,
    Shree
    ======================================================================================
    SELECT cr.cash_receipt_id
    ,cr.pay_from_customer
    ,cr.receipt_number
    ,cr.receipt_date
    ,cr.amount
    ,cust.account_number
    ,crh.gl_date
    ,cr.set_of_books_id
    ,sum(ra.amount_applied) amount_applied
    FROM AR_CASH_RECEIPTS_ALL cr
    ,AR_RECEIVABLE_APPLICATIONS_ALL ra
    ,hz_cust_accounts cust
    ,AR_CASH_RECEIPT_HISTORY_ALL crh
    ,GL_PERIOD_STATUSES gps
    ,FND_APPLICATION app
    WHERE cr.cash_receipt_id = ra.cash_receipt_id
    AND ra.status = 'UNAPP'
    AND cr.status <> 'REV'
    AND cust.cust_account_id = cr.pay_from_customer
    AND substr(cust.account_number,1,2) <> 'SI' -- Don't allocate Unapplied receipts FOR SI customers
    AND crh.cash_receipt_id = cr.cash_receipt_id
    AND app.application_id = gps.application_id
    AND app.application_short_name = 'AR'
    AND gps.period_name = 'May-07'
    AND crh.gl_date <= gps.end_date
    AND cr.receipt_number not like 'WH%'
    -- AND cust.customer_number = '0000079260001'
    GROUP BY cr.cash_receipt_id
    ,cr.pay_from_customer
    ,cr.receipt_number
    ,cr.receipt_date
    ,cr.amount
    ,cust.account_number
    ,crh.gl_date
    ,cr.set_of_books_id
    HAVING sum(ra.amount_applied) > 0;
    =========================================================================================
    Here is the explain plan in 10g r2 (10.2.0.2.0)
    PLAN_TABLE_OUTPUT
    Plan hash value: 2617075047
    | Id | Operation | Name | Rows | Bytes |TempSpc| Cost (%CPU)|
    | 0 | SELECT STATEMENT | | 92340 | 10M| | 513K (1)|
    |* 1 | FILTER | | | | | |
    | 2 | HASH GROUP BY | | 92340 | 10M| 35M| 513K (1)|
    | 3 | TABLE ACCESS BY INDEX ROWID | AR_RECEIVABLE_APPLICATIONS_ALL | 2 | 34 |
    | 4 | NESTED LOOPS | | 184K| 21M| | 510K (1)|
    |* 5 | HASH JOIN | | 99281 | 9M| 3296K| 176K (1)|
    |* 6 | TABLE ACCESS FULL | HZ_CUST_ACCOUNTS | 112K| 1976K| | 22563 (1)|
    |* 7 | HASH JOIN | | 412K| 33M| 25M| 151K (1)|
    | 8 | TABLE ACCESS BY INDEX ROWID | AR_CASH_RECEIPT_HISTORY_ALL | 332K| 4546K|
    | 9 | NESTED LOOPS | | 498K| 19M| | 26891 (1)|
    | 10 | NESTED LOOPS | | 2 | 54 | | 4 (0)|
    | 11 | TABLE ACCESS BY INDEX ROWID| FND_APPLICATION | 1 | 8 | | 1 (0)|
    |* 12 | INDEX UNIQUE SCAN | FND_APPLICATION_U3 | 1 | | | 0 (0)|
    | 13 | TABLE ACCESS BY INDEX ROWID| GL_PERIOD_STATUSES | 2 | 38 | | 3 (0)
    |* 14 | INDEX RANGE SCAN | GL_PERIOD_STATUSES_U1 | 1 | | | 2 (0)|
    |* 15 | INDEX RANGE SCAN | AR_CASH_RECEIPT_HISTORY_N2 | 332K| | | 1011 (1)
    PLAN_TABLE_OUTPUT
    |* 16 | TABLE ACCESS FULL | AR_CASH_RECEIPTS_ALL | 5492K| 235M| | 108K
    |* 17 | INDEX RANGE SCAN | AR_RECEIVABLE_APPLICATIONS_N1 | 4 | | | 2
    Predicate Information (identified by operation id):
    1 - filter(SUM("RA"."AMOUNT_APPLIED")>0)
    5 - access("CUST"."CUST_ACCOUNT_ID"="CR"."PAY_FROM_CUSTOMER")
    6 - filter(SUBSTR("CUST"."ACCOUNT_NUMBER",1,2)<>'SI')
    7 - access("CRH"."CASH_RECEIPT_ID"="CR"."CASH_RECEIPT_ID")
    12 - access("APP"."APPLICATION_SHORT_NAME"='AR')
    14 - access("APP"."APPLICATION_ID"="GPS"."APPLICATION_ID" AND "GPS"."PERIOD_NAME"='May-07')
    filter("GPS"."PERIOD_NAME"='May-07')
    15 - access("CRH"."GL_DATE"<="GPS"."END_DATE")
    16 - filter("CR"."STATUS"<>'REV' AND "CR"."RECEIPT_NUMBER" NOT LIKE 'WH%')
    17 - access("CR"."CASH_RECEIPT_ID"="RA"."CASH_RECEIPT_ID" AND "RA"."STATUS"='UNAPP')
    filter("RA"."CASH_RECEIPT_ID" IS NOT NULL)
    Here is the explain plan in 9i
    Execution Plan
    0 SELECT STATEMENT Optimizer=CHOOSE (Cost=445977 Card=78530 By
    tes=9423600)
    1 0 FILTER
    2 1 SORT (GROUP BY) (Cost=445977 Card=78530 Bytes=9423600)
    3 2 HASH JOIN (Cost=443717 Card=157060 Bytes=18847200)
    4 3 HASH JOIN (Cost=99563 Card=94747 Bytes=9758941)
    5 4 TABLE ACCESS (FULL) OF 'HZ_CUST_ACCOUNTS' (Cost=12
    286 Card=110061 Bytes=1981098)
    6 4 HASH JOIN (Cost=86232 Card=674761 Bytes=57354685)
    7 6 TABLE ACCESS (BY INDEX ROWID) OF 'AR_CASH_RECEIP
    T_HISTORY_ALL' (Cost=17532 Card=542304 Bytes=7592256)
    8 7 NESTED LOOPS (Cost=17536 Card=809791 Bytes=332
    01431)
    9 8 NESTED LOOPS (Cost=4 Card=1 Bytes=27)
    10 9 TABLE ACCESS (BY INDEX ROWID) OF 'FND_APPL
    ICATION' (Cost=1 Card=1 Bytes=8)
    11 10 INDEX (UNIQUE SCAN) OF 'FND_APPLICATION_
    U3' (UNIQUE)
    12 9 TABLE ACCESS (BY INDEX ROWID) OF 'GL_PERIO
    D_STATUSES' (Cost=3 Card=1 Bytes=19)
    13 12 INDEX (RANGE SCAN) OF 'GL_PERIOD_STATUSE
    S_U1' (UNIQUE) (Cost=2 Card=1)
    14 8 INDEX (RANGE SCAN) OF 'AR_CASH_RECEIPT_HISTO
    RY_N2' (NON-UNIQUE) (Cost=1740 Card=542304)
    15 6 TABLE ACCESS (FULL) OF 'AR_CASH_RECEIPTS_ALL' (C
    ost=60412 Card=8969141 Bytes=394642204)
    16 3 TABLE ACCESS (FULL) OF 'AR_RECEIVABLE_APPLICATIONS_A
    LL' (Cost=337109 Card=15613237 Bytes=265425029)

    Hi,
    The plan between 9i and 10g is pretty the same but the amount of data fetched has considerably increased. I guess the query was performing slow even in 9i.
    The AR_CASH_RECEIPT_HISTORY_ALL is presently having 332000 rows in 10g where as it was 17532 in 9i.
    AR_CASH_RECEIPT_HISTORY_N2 is now having 332,000 rows in 10g where as in 9i it had 1,740
    Try creating some indexes on
    AR_CASH_RECEIPTS_ALL
    hz_cust_accounts

  • Disappearing loops

    Hi everyone.
    I just bought a 15inch Macbook Pro and I've been having issues with the loops in Garageband.
    Let me take you through everything I have done.
    I got my Macbook home and I my old Mac mini to transfer its data onto my new Macbook Pro. My Mac mini was running iLife 06 very well.
    After the transfer had finished, and I started running Mac OS X for the first time on the laptop, I firstly updated the software, using "Software Updater". This included an update for Garageband.
    I open Garageband for the first time, and find that half of the loops are shaded grey, and won't playback. These are loops that I know exist because I used them on the previous version.
    I did some investigating and found the "Audio Loops" folder, "Library" -> "Audio". I discovered that there were two types of file in the folder; .aif and .caf.
    Half of the loops are just the .aif files. These are the loops that are coming up grey in Garageband.
    The other half of the loops have two copies of themselves, one .aif, the other .caf.
    I've tried to find a way to reinstall Garageband, but the Install Disc won't let me reinstall Garageband for some reason.
    How can I sort it out so that I can get these loops working in Garageband again.

    First of all: you shouldn't switch to a new computer by copying the files over by hand - you'll miss essential stuff that is in the Library folders. That's what the Migration Assistant is for.
    You can copy the Loops folder(s) over to your new system. If you have duplicate loops, keep the .caf version. After cleaning up, drag the folder(s) onto GB's Loop Browser, that should make those loops available in GB.

  • Logic 9 Pro won't load apple loops library or EXS24 Instruments!

    First off, specs:
    I'm running 9.1.8 on OSX 10.8.2
    I recently purchased a brand new 13" MacBook Pro (early this december) and used migration assistant to transfer my accounts, apps, and such from my old hard drive to my new one. Everything seems to be in its place and running fine, except for Logic 9 Pro.
    Whenever I load up a session featuring the EXS24 sampler, the loading bar will hang up upon loading the instrument, eventually turn into the spinning wheel of death, and become unresponsive. This also happens when I create a new session, open up a new instance of EXS24, and attempt to load a patch. All of the other stock Apple instruments load up just fine.
    I also noticed that Logic can't locate my Apple Loops Library, as every option in the loop browser is grayed out. I went into my library folder and found the EXS patches as well as the Apple loops, so I know they're there!
    Unfortunately, I don't have a Logic 9 CD (It came preinstalled on my previous computer), so to my knowledge reinstallation isn't an option, unless there's a way to reinstall Logic 9 without a CD and without losing my license. I've already tried the following to resolve this issue:
    - Reindexing spotlight
    - Reindexing loop browser
    - Disabling spotlight on the volume where the Logic files reside
    - Deleting the Logic 9 preference file
    If anyone has any insight or a way to reinstall Logic 9 Pro without a CD and without losing my license, it'd be greatly appreciated!

    Hi
    You probably need to re-index the Loop Library.
    In Logic 9 control click on the Loop Library and choose "Reindex"
    In Logic X use the Loops Menu at the top and again, choose "ReIndex".
    If that does not do it, you may need to manually remove the Index files
    CCT

  • GarageBand '11 won't download additional loops

    Hola,
    I have GarageBand '11, running on Mac OS X 10.6 Snow Leopard. The Apple loops have lots of greyed-out loops that need downloading. I click to activate one, it tells me I must download the loop now or later. I click Now, it goes to Software Update which tells me there is nothing left to download because I am completely up to date. I restart GarageBand, and still the loops are missing.
    It's very frustrating!
    In my Software Update download history it says the loops have already been downloaded, yet GarageBand still can't see them.
    Anybody got any idea what to do?
    Thanks,
    Graham

    I just ran into this and managed to solve it, but it was a pain.  Here's my situation:
    New iMac 27 inch
    Used Migration Assistant to transfer all user accounts, user files, and applications off of an older (late 2006) iMac that had been upgraded to Snow Leopard and iLife '09
    When trying to install the full set of Apple Loops and Software Instruments, I got the infamous "run Software Update but nothing ever downloads" problem.  After doing a lot of looking on these boards and elsewhere, I decided to uninstall GarageBand '11 and reinstall.  Uninstalling GarageBand '11 is a pain too; here's what I did:
    Deleted GarageBand from the Applications folder
    Deleted the /Library/Application Support/GarageBand folder
    Deleted the /Library/Audio/Apple Loops/Apple/Apple Loops for GarageBand folder
    Deleted anything related to GarageBand in /Library/Receipts
    And everything in ~/Library/Application Support/GarageBand (this is optional I think)
    I then reinstalled GarageBand from the second install disk that came with my new computer, fired up GarageBand, and downloaded the recommended update.  Finally, I opened up a new Loops project; in the Loops Browser, selected one of the missing loops; got the "run Software Update now or later" dialog, and said "Later"; quit GarageBand.  Finally, I opened Software Update, and lo and behold! the near-mythical GarageBand Loops and SW Instruments update was there, all 1.2 GB of it!
    I wish that Apple would come out with a patch for this, but in the meantime, if you're situation is similar to mine (upgraded GarageBand '11), you might try the deinstall / reinstall route and see if it works...

  • Loops not appearing in loop browser, why?

    I just got a new 20 inch iMac. I used Migration Assistant to help transfer the contents of my old G3 iMac DV SE from 2000. I was not running Garage Band on that computer.
    When I open up GarageBand on my new computer and then click on the loop browser, it gives me a dialog box that says something like: "No Apple Loops found. Please reinstall... yada yada." So, then I looked at a GarageBand article and saw that you can drag the loops into the browser. Well, I found a folder on my hard drive that said "Apple Loops" and it has 1,198 .aif files in it and dragged them into the browser. The progress bar said that it was importing the loops. Then when it was done, the loops were still not in the browser and it gave me the same "No Apple Loops found" box. What is going on?
    I had Final Cut Express preinstalled at the factory when I ordered this iMac. There is another folder with Soundtrack Loops. Could this be affecting GarageBand?
    I just want to get my loops into the loop browser. How do I do this?
    Also, I got discs that have the OS X reinstall on them, and Final Cut Express reinstall, but none with iLife. Was I supposed to get that disc too since it was preinstalled?

    http://www.bulletsandbones.com/GB/GBFAQ.html#noloops

Maybe you are looking for