Commit interval while removing duplicates

I am trying to remove duplicates from a table with over 10million records. Below query is working fine but it doesnt contain any COMMIT interval. I have to commit after every 20k or 30k records deletion for which IF loop is necessary.
Can anyone help me with it please ?
Query:
delete from
customer
where rowid in
(select rowid from
(select
rowid,
row_number()
over
(partition by custnbr order by custnbr) dup
from customer)
where dup > 1);

937851 wrote:
I am trying to remove duplicates from a table with over 10million records. Below query is working fine but it doesnt contain any COMMIT interval. I have to commit after every 20k or 30k records deletion for which IF loop is necessary. It is more efficient to delete all the rows and commit once
10M rows is a good number but modern systems should be able to support the delete. If you are worried about running out of rollback/undo you can (only if necessary!) batch deletes and perform periodic commits, something like (again, only if necessary)
delete from whatever
  where key in (result from subquery)
     and rownum <= 1000000;
delete from whatever
  where key in (result from subquery)
     and rownum <= 1000000;
delete from whatever
  where key in (result from subquery)
     and rownum <= 1000000;
. . .

Similar Messages

  • TS1717 iTunes crashed while removing duplicates

    I sorted my library by song title, and manually selected about 100 duplicate songs. I then right clicked and clicked "Delete." Then I chose to remove the songs from the library and disk. iTunes crashed. When restarted the library indicates missing files for those songs I wanted totally deleted. I checked some and the files are indeed deleted.
    Any advice on how to successfully remove these duplicates and how to fix my library would be appreciated.

    Lost & Found
    Create a playlist called Found, select everything in Music and drag it into the Found playlist (it may take some time to count the tracks that are to be dropped). Create a smart playlist called Lost matching All the rules Playlist is Music and Playlist is not Found. Your lost tracks will be in this playlist. You should be able to select all the tracks in this playlist with CTRL-A and then remove them from iTunes with CTRL-SHIFT-DELETE.
    Or use iTunes Folder Watch with its option to check for dead tracks on start up.
    As for tidying up duplicates, I've written a script called DeDuper which can help remove unwanted duplicates. See this  thread for background.
    tt2

  • There was an error while writing data back to the server: Failed to commit objects to server : Duplicate object name in the same folder.

    Post Author: dmface15
    CA Forum: Administration
    I am working in a new enviorment and i am trying to save a report to the Crystal Server via the CMC. I am uploading the report from the objects tab and attempting to save to a folder. The report has 1 static defined parameter and that's it. When i click submit to save the report i receive the following error message: "There was an error while writing data back to the server: Failed to commit objects to server : Duplicate object name in the same folder." There is not a anothe report within the folder with that name. What could be causing this error message and equally important, what is the solution.

    hte message you received about duplicate user probably means something hadn't fully updated yet. Once it did then it worked...
    Regards,
    Tim

  • XSLT to remove duplicates while concatinating

    My XML looks like folloing:
    <?xml version="1.0" encoding="utf-8" standalone="no"?> <BATCHES> <item> <Material>1000000079</Material> <Description>330 Bulk</Description> <Tank>T123</Tank> <Batch>2013225287</Batch> <Quantity>510</Quantity> </item> <item> <Material>1000000079</Material> <Description>330 Bulk</Description> <Tank>T123</Tank> <Batch>2013225301</Batch> <Quantity>520</Quantity> </item> <item> <Material>1000000196</Material> <Description>340R Bulk</Description> <Tank>T700</Tank> <Batch>1000188378</Batch> <Quantity>510</Quantity> </item> <item> <Material>1000002754</Material> <Description>43 Bulk</Description> <Tank>T515</Tank> <Batch>2013180125</Batch> <Quantity>300</Quantity> </item> <item> <Material>1000002754</Material> <Description>43 Bulk</Description> <Tank>T515</Tank> <Batch>2013203124</Batch> <Quantity>200</Quantity> </item> <item> <Material>1000002754</Material> <Description>43 Bulk</Description> <Tank>T515</Tank> <Batch>2013214839</Batch> <Quantity>700</Quantity> </item> <item> <Material>1000002754</Material> <Description>43 Bulk</Description> <Tank>T517</Tank> <Batch>2013214342</Batch> <Quantity>890</Quantity> </item> </BATCHES>
    My original XSLT look like this:
    <?xml version="1.0" encoding="utf-8" standalone="no"?> <xsl:stylesheet xmlns:xsl="http://www.w3.org/1999/XSL/Transform" version="1.0"> <xsl:output encoding="UTF-8" indent="yes" method="xml" version="1.0"/> <xsl:template match="/"> <Rowsets> <Rowset> <xsl:variable name="materials" select=".//item[Tank!='RECV' and Tank!='PARK'] "/> <xsl:for-each select="$materials"> <xsl:if test="generate-id(.)= generate-id($materials[Material=current()/Material])"> <Row> <Material> <xsl:value-of select="Material"/> </Material> <Description> <xsl:value-of select="Description"/> </Description> <Value> <xsl:for-each select="$materials[Material=current()/Material]/Tank"> <xsl:if test="node()"> <xsl:value-of select="concat(.,'||')"/> </xsl:if> </xsl:for-each> </Value> </Row> </xsl:if> </xsl:for-each> </Rowset> </Rowsets> </xsl:template> </xsl:stylesheet>
    The result of this XSLT looks like this:
    <?xml version="1.0" encoding="UTF-8" standalone="no"?> <Rowsets> <Rowset> <Row> <Material>1000000079</Material> <Description>330 Bulk</Description> <Value>T123||T123||</Value> </Row> <Row> <Material>1000000196</Material> <Description>340R Bulk</Description> <Value>T700||</Value> </Row> <Row> <Material>1000002754</Material> <Description>43 Bulk</Description> <Value>T515||T517||</Value> </Row> </Rowset> </Rowsets>
    I wanted to remove duplicate tanks while concatenating it in Value field. So I changed my XSLT to following:
    <?xml version="1.0" encoding="utf-8" standalone="no"?> <xsl:stylesheet xmlns:xsl="http://www.w3.org/1999/XSL/Transform" version="1.0"> <xsl:output encoding="UTF-8" indent="yes" method="xml" version="1.0"/> <xsl:template match="/"> <Rowsets> <Rowset> <xsl:variable name="materials" select=".//item[Tank!='RECV' and Tank!='PARK' and Quantity &gt; 500]"/> <xsl:for-each select="$materials"> <xsl:if test="generate-id(.)= generate-id($materials[Material=current()/Material])"> <Row> <Material> <xsl:value-of select="Material"/> </Material> <Description> <xsl:value-of select="Description"/> </Description> <Value> <xsl:for-each select="$materials[Material=current()/Material]/Tank[not(.=preceding::Tank)]"> <xsl:if test="node()"> <xsl:value-of select="concat(.,'||')"/> </xsl:if> </xsl:for-each> </Value> </Row> </xsl:if> </xsl:for-each> </Rowset> </Rowsets> </xsl:template> </xsl:stylesheet>
    My result now looks like this:
    <?xml version="1.0" encoding="UTF-8" standalone="no"?> <Rowsets> <Rowset> <Row> <Material>1000000079</Material> <Description>330 Bulk</Description> <Value>T123||</Value> </Row> <Row> <Material>1000000196</Material> <Description>340R Bulk</Description> <Value>T700||</Value> </Row> <Row> <Material>1000002754</Material> <Description>43 Bulk</Description> <Value>T517||</Value> </Row> </Rowset> </Rowsets>
    It removed the duplicate tank T123 for material 1000000079 but for material 1000002754, it even removed T515which should appear in Value field as its quantity is greater than 500 for following:
    <item> <Material>1000002754</Material> <Description>43 Bulk</Description> <Tank>T515</Tank> <Batch>2013214839</Batch> <Quantity>700</Quantity> </item>
    what am I doing wrong here?

    And if columns are NULLable:
    with sample_table as (
                          select '111' col1,'AAA' col2 from dual union all
                          select 'AAA','111' from dual union all
                          select '222','BBB' from dual union all
                          select '333','CCC' from dual union all
                          select '444',to_char(null) from dual union all
                          select to_char(null),'444' from dual union all
                          select to_char(null),to_char(null) from dual union all
                          select to_char(null),to_char(null) from dual
    select  least(col1,col2) col1,
            case
              when col1 is null then col2
              when col2 is null then col1
              else greatest(col1,col2)
            end
      from  sample_table s
      group by least(col1,col2),
               case
                 when col1 is null then col2
                 when col2 is null then col1
                 else greatest(col1,col2)
               end
      order by least(s.col1,s.col2),
               case
                 when s.col1 is null then s.col2
                 when s.col2 is null then s.col1
                 else greatest(s.col1,s.col2)
               end
    COL CAS
    111 AAA
    222 BBB
    333 CCC
        444
    SQL>  SY.

  • Remove duplicates while loading data from text file

    Hi,
    Data in text file (some times has duplicates) is being loaded into Oracle 9i database using Informatica. To improve performance, we would like to remove duplicates at the time of each load using Oracle procedure. Could you please help me with this?
    Thanks,
    Lakshmi

    No, our table doesn't have that. Most of the functionality is managed at the informatica level. Is there any other way? Thanks,

  • How to remove duplicate POP messages in Mail under Mavericks

    Hello
    I'm having some significant problems with Apple Mail, and despite having researched widely, I haven't been able to find a solution yet. Here is my tale of woe...
    My set up is that I use Mail with 5 separate accounts, downloading all messages using POP. I keep ALL my mail (even the trash) so that I have a permanent record of all conversations, going back more than 10 years - this is about 250,000 emails in around 250 subfolders. I'm running the latest version of Mavericks on a 2010 MBP. I have no intention of moving to IMAP - I need to keep a local archive of all my mail.
    A few months ago I was having problems with Mail being very slow to load and search emails. I decided to rebuild the mailboxes, but afterwards discovered that thousands of messages in the inbox of one of my accounts (and some of the sub-folders) had disappeared. Once I realised I tried to go back to a recent Time Machine back up, but, for whatever reason, that part of the back up hadn't worked properly and I was unable to recover the previous status. I left it while I decided what to do, and had time to deal with it...
    The last 2 years worth of emails on that account were still archived on gmail, so last week I decided to try redownloading the whole lot (something like 72,000 messages), and then planned to use one of the scripts I was aware of to remove the duplicates to the trash, and then remove any duplicates in the trash forever. That ought to mean I at least had an archive of the last two years of missing messages, even if I had no record of which ones had been flagged or replied to, etc.
    The downloading took about 24 hours to complete, and I now have almost 70,000 messages in the inbox of that account. I also rebuilt all the folders again, which doesn't seem to have resulted in any further loss of messages as far as I can tell. However, I hadn't figured on two things:
    1. Mail 'hides' duplicate messages, meaning that all of the messages I've downloaded are showing as unread, and I can't differentiate the ones I already had in my Inbox, and the new downloads. When I click on those apparently unread messages I can see if they have been replied to or forwarded, etc, but there's no obvious way for me to remove the ones I have already dealt with to the trash, without going through them all manually, which is clearly impossible.
    2. The scripts I'd found for removing duplicates don't work. Andreas Amann's Remove Duplicates script doesn't work under Mavericks, and he has abandoned the project. I've also tried the remove-duplicate-messages.scpt made by Jolly Roger (http://jollyroger.kicks-***.org/software/), and while it *sometimes* works on individual subfolders on my Mac (but as far as I can tell removes the duplicate in that folder, rather than the newly downloaded version), mostly it doesn't work at all - it creates a 'Remove Duplicate Messages' folder on my desktop, a log inside it and a folder for removed messages, but nothing appears in the duplicates folder.
    So, I'm left in a position where I have 70,000 apparently unread messages in my Inbox, a massively bloated Mail library (which has pretty much doubled in size, because of the 'hidden' messages), a slow and unresponsive Mail program. I've come to the conclusion that there must be some corrupted email somewhere, which probably caused the original email haemorrhage, and may still be causing the inability to remove duplicates. Mail is so slow as to be almost unuseable.
    I figure I have a number of options:
    1. I could live with the situation and just archive most of the Mail in my inbox, with the side effect that there will be a bunch of messages I have never replied to that are missed.
    2. I could abandon the last week's efforts and revert to the version of Mail I was using a week ago, and then redownload the recent emails from my various accounts. That would still leave me without those thousands of emails I lost on my local machine.
    3. I could find another way to deal with this. Can I get the remove duplicates script working? Should I revert to the version of Mail from a week ago, download ALL the messages again, but do it in a way that allows me to find the duplicates and remove them? Should I move to another email program altogether (which would presumably be massively disruptive to my work!)
    Anyway, apologies for the essay length of this request, and thanks in advance for any help you may be able to offer!

    hi Eric
    thanks for the tip, but I don't think that solves my problem, which is that Mail apparently hides the duplicates, so they're still there, you just can't differentiate or remove them. I actually *need* the duplicates there so I know I've got everything, but then I need to be able to move the ones I know I've dealt with from my inbox to an archive/trash, and/or remove them completely so Mail isn't totally bloated

  • How to remove duplicate messages from Mail in Mavericks (POP)

    Hello
    I'm having some significant problems with Apple Mail, and despite having researched widely, I haven't been able to find a solution yet. Here is my tale of woe...
    My set up is that I use Mail with 5 separate accounts, downloading all messages using POP. I keep all my mail so that I have a permanent record of all conversations, going back more than 10 years - this is about 250,000 emails. I'm running the latest version of Mavericks on a 2010 MBP. I have no intention of moving to IMAP - I need to keep a local archive of all my mail.
    A few months ago I was having problems with Mail being very slow to load and search emails. I decided to rebuild the mailboxes, but afterwards discovered that thousands of messages in the inbox of one of my accounts (and some of the sub-folders) had disappeared. Once I realised I tried to go back to a recent Time Machine back up, but, for whatever reason, that part of the back up hadn't worked properly and I was unable to recover the previous status. Ileft it while I decided what to do, and had time to deal with it.
    The last 2 years worth of emails on that account were still archived on gmail, so last week I decided to try redownloading the whole lot (something like 72,000 messages), and then uplanned to use one of the scripts I was aware of to remove the duplicates to the trash, and then remove any duplicates in the trash forever. That ought to mean I at least had an archive of the last two years of missing messages, even if I had no record of which ones had been flagged or replied to, etc.
    The downloading took about 24 hours to complete, and I now have almost 70,000 messages in the inbox of that account. However, I hadn't figured on two things:
    1. Mail 'hides' duplicate messages, meaning that all of the messages I've downloaded are showing as unread, and I can't differentiate the ones I already had in my Inbox, and the new downloads. When I click on those apparently unread messages I can see if they have been replied to or forwarded, etc, but there's no obvious way for me to remove the ones I have already dealt with to the trash, without going through them all manually, which is clearly impossible.
    2. The scripts I'd found for removing duplicates don't work. Andreas Amann's Remove Duplicates script doesn't work under Mavericks, and he has abandoned the project. I've also tried the remove-duplicate-messages.scpt made by Jolly Roger (http://jollyroger.kicks-***.org/software/), and while it *sometimes* works on individual subfolders on my Mac (but as far as I can tell removes the duplicate in that folder, rather than the newly downloaded version), mostly it doesn't work at all - it creates a 'Remove Duplicate Messages' folder on my desktop, a log inside it and a folder for removed messages, but nothing appears in the duplicates folder.
    So, I'm left in a position where I have 70,000 apparently unread messages in my Inbox, a massively bloated Mail library (which has pretty much doubled in size), a slow and unresponsive Mail program. I figure I have a number of options. Either I could abandon the last week's efforts and revert to the version of

    Apologies, this was posted here in error. Please find updated post here: https://discussions.apple.com/message/27261965#27261965

  • Help needed in removing duplicate items of list box  in java

    How to remove duplicate items of list box while dynamically inserting (on-click event)
    It is not identifying duplicate data
    Variable name is HP_G1
    HP_dmg1 = (DefaultListModel) HP_G1.getModel();
    int a = HP_G1.getModel().getSize();
    System.out.println("HP list no--------> "+a);
    if(a!=0)
    for (int j=0; j<a; j++)
    String item1 = String.valueOf(HP_List.getModel().getElementAt(j));
    System.out.println("HP list added--------> "+item1);
    if(HP_dmg1.equals(item1)){
    HP_dmg1.remove(j);
    else
    HP_dmg1.addElement(GPL);
    }

    Your code is unreadable, so I'll ignore it. In the future please press the message editor's CODE button to format code.
    As to your problem, to the point you normally use a Set instead of List when you don't want duplicates in a collection.

  • Help in SQL database remove duplicate fields

    I have different tables like Employee, mentor_offer, department and criteria.  Employee has all the employee details and department has the id and name of all the departments, Criteria has id and name of all the skills. Mentor_offer has the email
    id and the skill for that particular email id.
    The stored procedure that i am using is 
    SELECT distinct e.firstName + ' ' + e.lastName as name, e.email, e.geo,c.name as skill, d.name as department, e.addedDate, e.modifiedDate 
    from employee e 
    inner join mentor_offer m 
    on e.email=m.email
    inner join department d
    on e.department = d.id
    inner join criteria c 
    on m.criteria_id = c.id
    order by e.addedDate desc
    At present I get the result like this:
    Name|Email|Location|Department|Skill|addedDate|mod ifiedDate
    Arun Rao|[email protected]|NA| Marketing|Planning & Organizing| 3/17/2014 1:01:21 PM|3/20/2014 2:37:50 PM
    Arun Rao|[email protected]|NA| Marketing|Technical Skills| 3/17/2014 1:01:21 PM|3/20/2014 2:37:50 PM
    Binay|[email protected]|NA|Sales|Leadership|3/17/2014 1:01:21 PM|3/20/2014 2:37:50 PM
    Binay|[email protected]|NA|Sales|Technical Skills|3/17/2014 1:01:21 PM|3/20/2014 2:37:50 PM
    Expected Result:
     I would like to add the skills in a single row with comma separated and remove the duplicate entries of Name, Email ,Location, Department. 
    Name|Email|Location|Department|Skill|addedDate|mod ifiedDate
    Arun Rao|[email protected]|NA| Marketing|Planning & Organizing, Technical Skills|3/17/2014 1:01:21 PM|3/20/2014 2:37:50 PM
    Binay|[email protected]|NA|Sales|Leadership, Technical Skills|3/17/2014 1:01:21 PM|3/20/2014 2:37:50 PM
    Please help.

    Try this: 
    with CTE
    as (
    select distinct
    e.firstName + ' ' + e.lastName as name
    , e.email
    , e.geo as Location
    , c.name as skill
    , d.name as department
    , e.addedDate
    , e.modifiedDate
    from
    employee e inner join mentor_offer m
    on e.email = m.email
    inner join department d
    on e.department = d.id
    inner join criteria c
    on m.criteria_id = c.id
    select
    name
    , email
    , Location
    , department
    Skill = stuff((
    select
    ', ' + skill
    from cte b
    where
    c.name = b.name
    and c.email = b.email
    and c.Location = b.Location
    and c.department = b.department
    for
    xml path('')), 1, 2, '')
    max(addedDate) as addedDate
    max(modifiedDate) as modifiedDate
    from cte c
    group by
    name
    , email
    , Location
    , department
    order by
    addedDate desc;
    not tested/.
    Regards Harsh

  • Merge and remove duplicates

    I would like to use a utility or 3rd party app to go through my two 90gb libraries of photos (iPhoto libraries) and merge them while also removing duplicates and potentially trimming it (the library was originally but in iPhoto 08 or 07, probably)
    How can I merge these libraries and soft out duplicates? The majority of the libraries are identical contents FYI, but there might be a thousand unique.
    Thanks experts!

    If you have Aperture 3.3 or later and iPhoto 9.3 or later you can merge libraries with Aperture. You can then use one of the many duplicate finders to sort out the problem.
    Otherwise the only way to merge Libraries while avoiding duplicates is with the paid ($20) version of iPhoto Library Manager

  • HashSet doesnt remove duplicates

    Hello
    I am trying to remove duplicates from a list using this code
    Set set = new HashSet();
    ListIterator li = list.listIterator();
    while (li.hasNext())
                   A pt = (A) li.next();
                   duplicate = pt.getID(); //duplicate is declared globally
                    if (!set.add(pt))
                        System.out.println("Duplicate found: " + pt.getID());
               }but it isnt removing duplicates. I read somewhere that I should override the equals() and hashCode() methods. But I am a bit confused as to how to write my equals method. The list contains objects of class A, which has many getters and setters. The elements are considered to be duplicate when getID() getter returns same value for two objects. Can some one tell me how should i go about writing the equals method. And why the default equals method for objects isnt working in this case?
    This is how i tried to write my equals() method but it isnt working
    public boolean equals(Object o)
           if (o == this)
                 return true;
            if (!(o instanceof A))
                 return false;
            A pn =  (A) o;
            return pn.getID() == duplicate; // duplicate has been set before calling add
         }

    Sorry if I got confusing about what I was trying to say. I was actually trying to expand on your earlier post, where you mentioned that the class needs the proper implementation of equals() had hashCode().
    I went over this with one of my programmers fairly recently. He was having a hard time understanding why he was getting a false when the contents of the object were the same. Of course he did not override the equals() method and was doing an object comparison, and not one on the contents of the objects.
    Maybe the following examples will better illustrate what I was trying to say.
    Without equals:
    public class EqualityExample {
         public EqualityExample()
          * @param args
         public static void main(String[] args) {
              EqualityExample ee = new EqualityExample();
              // TODO Auto-generated method stub
              DataClass testObject1 = ee.new DataClass("A String");
              DataClass testObject2 = ee.new DataClass("A String");
              if (testObject1.equals(testObject2))
                   System.out.println("are equal");
              else
                   System.out.println("not equal");
         public class DataClass
              private String someVal;
              public DataClass(String someVal)
                   this.someVal = someVal;
    }With equals:
    public class EqualityExample {
         public EqualityExample()
          * @param args
         public static void main(String[] args) {
              EqualityExample ee = new EqualityExample();
              // TODO Auto-generated method stub
              DataClass testObject1 = ee.new DataClass("A String");
              DataClass testObject2 = ee.new DataClass("A String");
              if (testObject1.equals(testObject2))
                   System.out.println("are equal");
              else
                   System.out.println("not equal");
         public class DataClass
              private String someVal;
              public DataClass(String someVal)
                   this.someVal = someVal;
              public boolean equals(DataClass object)
                   if (object.someVal == this.someVal)
                        return true;
                   else
                        return false;
    }

  • How to remove duplicates

    Hi
    i am removing duplicate records while importing bulk data into the table...I am checking for some columns...when they are same, i am removing the old records...i have used the following code to remove duplicates...
    execute immediate 'DELETE FROM test1 WHERE ROWID IN (SELECT ROWID FROM (SELECT ROWID,ROW_NUMBER() OVER (PARTITION BY c1,c2 ORDER BY 1) row_no FROM test1)WHERE row_no > 1)';
    here i check c1 and c2 columns...if they are same the old records are to be deleted...but in this code, the new records are deleted..can anyone say how to remove old duplicate records?
    Vally

    Hi
    i am removing duplicate records while importing
    bulk data into the tableWhat you mean by using "while"?
    During the process of importing(read inserting) - you want to delete duplicate records?
    As you say in the following you have C1 and C2 - using both of them - you find duplicates.
    I deem you have other columns besides C1 and C2. And these columns have different fileds in NEW record and OLD record - then why don't you use UPDATE statement?
    ...I am checking for some
    columns...when they are same, i am removing the old
    records...i have used the following code to remove
    duplicates...you should clarify on what criteria you separate old records from new records and place this condition in your query.
    E.g. you have a field DATE_OF_ENTRY
    and the latest one is the new record which shouldn't be deleted
    then you would be able to put it into your delete statement:
    DELETE FROM test1
    WHERE ROWID IN (SELECT ROWID
                       FROM (SELECT ROWID,
                                    ROW_NUMBER() OVER(PARTITION BY c1, c2 ORDER BY DATE_OF_ENTRY desc) row_no
                               FROM test1)
                      WHERE row_no > 1)

  • Removing duplicate emails in Mail

    While attempting to back up my Gmail -- which is at capacity right no -- I accidentally reset the POP setting and downloaded all my mail again.
    So now I have 160000+ emails in my mailbox in Mail. I would like to weed out the dupes, and after a bit of Googling, found Mail Scripts (http://www.versiontracker.com/dyn/moreinfo/macosx/16217) and started running the Remove Duplicates script.
    Everyone said it was slow, but after leaving it running all night it had found only 119 duplicates! Is there any other program that will remove the extra mail, or there some way to shorten the process? It seems like my mailbox is just too big for this script to handle.

    Just a thought, but take a look at your Mail Menu "View", select columns, then make sure it has both Date Received and Date Sent checked. Then compare duped messages and see if the date/time is the same on both. I haven't had this problem, but if the Date Received is different on the second copy of the duped email, you might be able to sort on Date Received and mass-delete the dupes...
    beadman

  • HT2905 remove duplicates in Itunes 11. How???

    I need to remove duplicate music in Itunes 11, but don't have a duplicate option under File. Any suggestions?

    look up where it says File Edit View Controls Store Help
    click View
    then click Show Duplicate Items
    then while holding Ctrl click on each song you want to delete
    right click, delete

  • I want to remove duplicate vendor number

    I want to remove duplicate vendor number,  should I put deletion flag or block them ?
    Also in some cases there are some dependancies (open PO, pending invoice, payment due), can somebody suggest safe way ?

    I would recommend to set both indicators.
    Block will restrict the usage, while deletion indicator makes the vendor going to the archive file/removing it from the table with the next archiving run.
    You have to decide what you want to do with the open business cases.
    You either convert them to the correct vendor directly, or you finish it during the next month with standard business process.
    Blocking of vendor master may have some impact to open business cases.

Maybe you are looking for

  • Final Cut Pro 6 now quits on me everytime I try to open it!

    Please people, I really need some help! I am an editor and I have been using Final Cut Pro 6 for a couple of years now. I'm constantly working on projects so I really need Final Cut up and running almost all the time! Recently (today actually) I trie

  • Shortcut to Web App

    Is it possible to create a shortcut with Nal on the desk top to a Webpage/ WebApp or put automatically in the users bookmarks?

  • Is there a way I can tell if the unlock worked?

    I just followed the instructions to unlock an old iPhone 3gs that I want to take on vacation overseas. I didn't get any sort of message when I was done. I don't want to be stuck overseas figuring out an unlock w/o my normal gear. Is there a way to te

  • Hiding the Comments in SE38 Display Mode

    Hi Folks, Is it possible to Hide the comments when a program is seen in uneditable mode either in SE38 or SE80.I am using 4.7. If not in 4.7 kindly let me know if any of the upcoming versions have this facility. Thanks, K.Kiran.

  • Cannot prevent authenticated users from creating a blog on "My Page"

    I have a brand new Snow Leopard (10.6.1) 2.26 Ghz quad core Xserve with 12Gb RAM that will be used for web collaboration services. I've currently set up Wiki and Blog services with a group membership to allow creating wikis/blogs. The reason for this