How to get most recent consecutive records with the same value

Hi,
Can someone please help me to solve my problem?
Below is my table
Prod_Code-----Purchase_date---Inv_number-------Inv_amt
MS_AV16850------18/01/2005-----------6575947----------------7.93
MS_AV16850------22/07/2005-----------7034012----------------51.82
MS_AV16850------04/01/2006-----------8719618----------------51.82
MS_AV16850------20/06/2006-----------9515864----------------104.69
MS_AV16850------16/04/2007-----------10353759----------------189.29
MS_AV16850------30/05/2007-----------10689899----------------189.29
MS_AV16850------06/01/2008-----------1653821----------------65.49
MS_AV16850------22/02/2009-----------10866311----------------189.29
I want my query to show the rows that has most recent purchase dates with same amount in consecutive purchase date.
So from the table above, the query should display:
Prod_Code-----Purchase_date---Inv_number-------Inv_amt
MS_AV16850------16/04/2007-----------10353759----------------189.29
MS_AV16850------30/05/2007-----------10689899----------------189.29
It should not get
MS_AV16850------16/04/2007-----------10353759----------------189.29
MS_AV16850------30/05/2007-----------10689899----------------189.29
MS_AV16850------22/02/2009-----------10866311----------------189.29
because inv_number 10866311 has a prvevious inv_amount of 65.49.
and not get this
MS_AV16850------22/07/2005-----------7034012----------------51.82
MS_AV16850------04/01/2006-----------8719618----------------51.82
because they are not the most recent purchase date even if they have the same inv_amount.
Thanks in advance.

Hi,
You're right; thanks for catching my mistake.
I changed the WHERE clause of the main query (including subquery there) to deal with that situation:
WITH     got_grp          AS
     SELECT      x.*
     ,      ROW_NUMBER () OVER ( ORDER BY      purchase_date )
            - ROW_NUMBER () OVER ( PARTITION BY  inv_amt
                                         ORDER BY         purchase_date )     AS grp
--     ,      ROW_NUMBER () OVER ( ORDER BY      purchase_date )     AS r1
--     ,        ROW_NUMBER () OVER ( PARTITION BY  inv_amt
--                                         ORDER BY         purchase_date )     AS r2
     FROM     table_x     x
SELECT     *     -- Or list all columns except grp
FROM     got_grp
WHERE     (inv_amt, grp)  IN  (
                           SELECT    MAX (inv_amt) KEEP (DENSE_RANK LAST ORDER BY MAX (purchase_date))
                    ,       MAX (MAX (grp))
                            FROM      got_grp
                            GROUP BY  grp
                            ,       inv_amt
                            HAVING    COUNT (*)     > 1
;Thanks, too, for posting the sample data. Apparantly, you're more interested in solving this problem than OP is.
Edited by: Frank Kulash on Nov 22, 2010 1:36 PM
The r1 and r2 columns are not needed for the solution. You may want to display them, just to help see how the query works.

Similar Messages

  • How to get list of active users with the details like samaccountname, name, department, job tittle, email in active directoy?

    how to get list of active users with the details like samaccountname, name, department, job tittle, email in active directoy?

    You can use third party software True Last Logon 2.9.You can export the file in excel for report creation.You can use the trial version this will achieve what you are looking for.
    True Last Logon displays the following Active Directory information:
    --Users real name and logon name
    --Detailed account status
    --Last Logon Date & Time
    --Last Logon Timestamp (Replicated value)
    --Account Expiry Date & Time
    --Enabled or Disabled Account
    --Locked Accounts
    --Password Expires
    --Password Last Set Date & Time
    --Logon Count
    --Bad Password Count
    --Expiry Date
    --You can also query for any other attribute (Example: Description, telephone Number, custom attibutes etc)
    Refer the below link for trial version:
    http://www.dovestones.com/products/True_Last_Logon.asp
    Best Regards,
    Sandesh Dubey.
    MCSE|MCSA:Messaging|MCTS|MCITP:Enterprise Adminitrator |
    My Blog
    Disclaimer: This posting is provided "AS IS" with no warranties or guarantees , and confers no rights.

  • How to get online apps to work with the Verizon DSL modem firewall set at "Medium" level, not "Low?"

    How to get online apps to work with the Verizon DSL modem firewall set at “Medium” level? Xbox 360 Live, FTP, and the Windows sntp Time checker native to Windows XP WON’T WORK unless the Verizon firewall is first reset down to “Low.” Then all works well, but I then risk low firewall protection. Setup: · Westell 6100 DSL modem (Software Version: VER:4.04.03.00 Transceiver Revision: 7.2.3.0 Model Name: C90-610015-06), · D-Link DIR-655 wireless router (Hardware Version: A3, Firmware Version: 1.21), For Xbox 360 Live the “canned” port forwarding rule for provided in Verizon’s drop-down list for the Westell 6100 modem forwards ports 88 and 3074 for both TCP and UDP. This does not meet Microsoft’s forwarding recommendations for Xbox 360 Live stated at http://support.microsoft.com/kb/908874. Instead, for the Verizon modem (Westell) I made a new port forwarding rule to comply with those Microsoft instructions. It forwards ports 53, 80 and 3074 for TCP and ports 53, 88 and 3074 for UDP. These same ports are also forwarded in the D-Link router configuration. For FTP, port 21 is also forwarded for TCP and UDP in both modem and router. But they don’t work until I first cripple the Verizon firewall down to “Low.” How to get them to work with the Verizon firewall up to Medium?

    While I do not know the answer to your question, BUT since you have this modem and another router - you could just follow.
    http://www.dslreports.com/faq/13600
    ^^
    If you are the original poster (OP) and your issue is solved, please remember to click the "Solution?" button so that others can more easily find it. If anyone has been helpful to you, please show your appreciation by clicking the "Kudos" button.

  • How can I get "IMAQ Line Fit" stable (with the same points)

    How can I get "IMAQ Line Fit" stable (with the same points)
    I my vision applications I use a lot off times "IMAQ Line Fit"
    some times the results are not Stable.
    See vi and word document attached.
    When I take the same points but in a other order (sort)
    then suddenly all points are used.
    In this case I was looking for a edge in vertical direction (Y).
    I found the points but "IMAQ Line Fit" did not function well.
    I tried changing all settings like pixel radius and minimum score and
    also threshold and so on. But no result.
    Then I sorted the points reversed and that worked. See vi.
    I use labview Labview 8.6.0 and IMAQ machine vision.
    Maybe this problem is solved in a newer version.
    Please let me know.
    Thanks
    Attachments:
    TEST-imaq-line-fit-86.vi ‏169 KB
    IMAQ-Fit-Line-Problem-86.doc ‏230 KB

    Thank You for your reaction Hossein,
    This "one application" uses "Line fit" about 20 000 000 times a day.
    about 200 will go wrong. With your setting the the "Line fit" for this occasion will work but others will go wrong.
    A algorithm should be:
    1e Stable, when using the same points and settings the result MUST be the same. How the points are sorted should not matter.
    2e Accuracy
    3e Speed
    4e Testing. Is very difficult but al least close points and the same points and sorting ... should be done
    NI now as made a " IMAQ Line Fit" Problem CAR (Corrective Action request) 298016.
    so in the future there will be a update.
    In the mean while a made a line fit which uses Bisquare option from Linear Fit.vi :
      from C:\Program Files\National Instruments\LabVIEW 8.6\vi.lib\analysis\6fits.llb
    This is a temporary fix because "IMAQ Fit Line VI" is not so stable.
    Because outliers should be rejected only Bisquare does this reasonable.
    This program does NOT USE "Minimum Score" and "Pixel Radius"
    Valid fit = When there are enough points within Outlier_Distance.
    Algorithm steps:
    - Determine if line is more horizontal then vertical, Compare DeltaX with DeltaY
    - If DeltaX <= DeltaY then SWAP X Y and later Swap back
    - Sort Points at on X value
    - First : Bisquare Linear Fit
    - Calculate distances to the found line for all points
    - Inliers: Points within Outlier_distance
    - Outliers: Points outside Outlier_Distance
    - Only use Inlier Points
    - Check again if DeltaX < DeltaY then SWAP X Y later SWAP back again
    - Second : Bisquare Linear Fit only with the Inlier Points
    - Results are from the second fit
    Warning : This Program is Not so accurate and Not so fast
    See attachments it is in labview 8.6
    If some has a improvement please let me know.
    Thanks
    Attachments:
    svi_fit_line_Bsquare_and_draw.vi ‏18 KB
    Fit_Line_Bisquare_Linear.vi ‏40 KB

  • Send emails from a csv file (grouping records with the same email addresses)

    Writing a script to send emails from a csv which contains record details and email addresses using the Send-Mailmessage cmdlet.
    That part is no trouble, however, the csv file contain records that can have the same email address.
    Is there a way/method I could send the records with the same email address once instead of sending individual email messages?
    Thanks
    Data in the csv file
    Record number, description, email
    1234, Test 1, [email protected]
    5678, Test 2, [email protected]
    1245, Test 3, [email protected]
    4578, Test 4, [email protected]
    $data = Import-csv c:\records.csv
    ForEach($address in $data)
    Send-mailmessage -To $address.emails -from [email protected] -subject "List of records" -Body $data -SmtpServer 192.168.1.1

    The following code does what you are looking for:
    $data = Import-csv c:\records.csv$mx = "192.168.1.1"
    $subject = "List of records"
    $from = "[email protected]"
    $data| Group-Object email|Select-Object Name, @{n='msg'; e={$_.Group| Select-Object -Property "record number", description|ConvertTo-Csv -notypeinformation|Out-String}}|
    ForEach-Object {Send-MailMessage -to $_.Name -Body $_.msg -SmtpServer $mx -Subject $subject -From $from}
    I also recommend no forcing everything into one line.  "One-liner" means "one pipeline" not putting all code on one unreadable line.
    Your code should look like this:
    $data|
    Group-Object email|
    Select-Object Name, @{
    n='msg'; e={
    $_.Group| Select-Object -Property "record number", description|
    ConvertTo-Csv -notypeinformation|Out-String
    }|
    ForEach-Object {
    Send-MailMessage -to $_.Name -Body $_.msg -SmtpServer $mx -Subject $subject -From $from
    Now we can see the code and see that you are still unnecessarily converting back and forth.
    To get the group look at how I did it.
    $body=$_.Group | Format-Table |Out-String
    Isn't that much easier?
    Once you master the pipeline these things will become second nature.  In PowerShell it is not necessary to write lots of code most of the time.
    ¯\_(ツ)_/¯

  • DTP error: No Message: Filter Out New Records with the Same Key

    Hello SDN,
    I have looked up solutions for this problem, but didnt find anything suitable. I am getting this error in multiple places: when i load pure master data into IOs like 0VENDOR, 0PLANT etc. or when I load DSOs like 0FIAP_O03.
    When I checked the load in sdn, one of the solutions was to apply note: 1489178 , but in this case, the user had nothing selected in his 'semantic key' list. Whereas, in my case.. 0PLANT is checked ...eventhough the other fields are disabled. So, I think my DTP is accurate.. as 'plant' is the only key field here.  But it still fails .. and so does 0VENDOR load. Moreover... the DSO load is crucial for our reports in Prod and its failure is a bigger issue for us.
    Has anyone seen this problem? Is there another solution other than the above note?
    Regards.

    Hello Arvind,
    If I select that option, will it delete one record if it finds 2 records with the same semantic key in 1 data packet. Also, will it delete a record if it finds that there is already a record existing in the data target with the same semantic key.
    I think, both of the above scenarios will result in incorrect data. From my understanding, 'Handle Duplicate Keys' is mostly used to filter out master data.. as in some instances.. you might get duplicate entries.
    But, in my case, the same error is happening in the case of data load from 0fi_ap_4 --> 0fiap_o03. What happens is... the overall status of the DTP becomes green.. but when you look at details section, the process stops at this message which is in yellow. Because of this, the DTP status in the PC remains yellow.. and finally turns red.. and the PC turns red.
    From my knowledge, it appears more like a bug. I was hoping to find a SAP NOTE which has some manual instructions that can resolve the issue, since the above NOTE asks for a support pack upgrade.. which would need a whole lotta testing.
    Regards.

  • I currently have aT-Mobile sim, and therefore, a UK telephone number. Can I get a sim from EE with the same number to use in my iphone 5?

    I currently have aT-Mobile sim, and therefore, a UK telephone number. Can I get a sim from EE with the same number to use in my iphone 5?

    The easiest way to find out what carrier it's locked to would be to call the store where you purchased the phone.
    The phone can be unlocked from any location.  However, most carrier have restrictions about unlocking and you have to contact them for details.

  • Reducing time while Filtering Out New Records with the Same Key

    Hi Experts,
    I have an issue .. I am trying to load data into 0MATERIAL . While loading via DTP the step u201CFilter Out New Records with the Same Keyu201D   takes much time for processing . For each datapackage it is taking around 30 minutes to filter out new records with the same key .  While for other master data it hardly takes 5 minutes.
    Any pointers how to reduce the timeu2026
    Thanks in advance
    Sam

    Hello,
    No there is no need to do that change then.
    Can you tell me are you doing a full load daily to 0MATERIAL object.
    why dont you use delta.
    Also if the DTP is of type FULL, then are you deleting the previous PSA requests.
    Maybe you can ask the basis guys to check with a trace what is really happening on the DB side during this 30 mins and then maybe we can find the needed fix.
    Regards,
    Shashank

  • Record filtered in advance as error records with the same key exist

    Dear Masters
       I am getting below error for master data (0MATERIAL_ATTR), while loading data from PSA to Info object through DTP (I have set the Handle duliocate record also). When I checked that error data in Error Stack, I am getting below error for all records(100)
    "Record filtered in advance as error records with the same key exist"
    Could u please help me to resolve this issue
    Thanks in Advance
    Raja

    Hi
    Thanks for reply
    I have loaded the Delta data in PSA. In PSA data was loaded successfully. I got the failure only when I loaded the from PSA to Maste data
    Thanks in advance
    Raja

  • How can I set up two iPhones with the same contact, photos, music etc... but a different sim card and phone number. They will both be my phones, but one will be for use in other countries.

    How can I set up two iPhones with the same contact, photos, music etc... but a different sim card and phone number. They will both be my phones, but one will be for use in other countries.

    Phone A = phone with information you want duplicated
    Phone B = phone that you want to copy from A
    Backup Phone A.
    Wipe Phone B using these instructions. What to do before selling or giving away your iPhone, iPad, or iPod touch - Apple Support
    Phone B should be on the 'Hello' screen. Do not swipe it. Launch iTunes. Plug in Phone B.
    When iTunes asks if you want to set up as new phone or restore from backup, choose restore from backup of Phone A.
    After this setup, any changes to one phone (i.e. adding a contact, downloading an app) will have to be manually duplicated on the second phone, if that is your desire). Making changes to one phone will not affect the other after setup.

  • How to Set JFilechooser to show file with the same extension in the filetyp

    How to Set JFilechooser to show file with the same extension in the filetypes??
    help

    I also don't understand, but if you mean how to display only files with a certain extension in your JFileChooser then you should have a look at the JFileChooser API:
    http://java.sun.com/j2se/1.5.0/docs/api/javax/swing/JFileChooser.html
    The introduction shows an example of how to do that.

  • How can i find my lose iphone with the same apple id i use now

    how can i find my lose iphone with the same apple id i use now

    If you set it up on your iPhone before you lost it then you can sign into your iCloud account on a computer and attempt to locate it.

  • How to get my custom controller updated with the global custom controller?

    Hi all,
    I'm new to CRM Web UI and need some advice from the expert. Currently I'm working on component ICCMP_BT_DATES and noticed something weird with this component. When this component is first launch it display the dates of a service ticket correctly. However when I navigate to another screen, save a new ticket and back, the dates are not reflected. When I went in and debug the component, I noticed that the context is still tied to the previous ticket. I think the custom controller is not updated with the latest from the global custom controller.
    My question is how do I get my custom controller updated with the latest.
    Regards,
    Ricky

    You have to bind your custom controllers context node to the event NEW_FOCUS of the collection wrapper on the global custom controller.
    Best place to do this might be the CONNECT_NODES of the context of your custom controller.
    Get the global CuCo with GET_CUSTOM_CONTROLLER() and then the appropriate context node. Now:
    SET HANDLER yourMethod for lr_global_cuco->typed_context->thecontextnode->collection_wrapper activation iv_activate.
    Of course you have to implement a method similar to ON_NEW_FOCUS as it is on many other nodes.
    cheers Carsten

  • One-to-many selfjoin removing records with the same ranking or with a substitute

    Sorry for my bad choice of discussion title, feel free to suggest me a more pertinent one
    I've rewritten post for clarity and following the FAQ.
    DB Version
    I'm using Oracle Enterprise 10g 10.2.0.1.0 64bit
    Tables involved
    CREATE TABLE wrhwr (
    wr_id INTEGER PRIMARY KEY,
    eq_id VARCHAR2(50) NULL,
    date_completed DATE NULL,
    status VARCHAR2(20) NOT NULL,
    pmp_id VARCHAR2(20) NOT NULL,
    description VARCHAR2(20) NULL);
    Sample data
    INSERT into wrhwr  VALUES (1,'MI-EXT-0001',date'2013-07-16','Com','VER-EXC','Revisione')
    INSERT into wrhwr VALUES  (2,'MI-EXT-0001',date'2013-07-01','Com','VER-EXC','Verifica')
    INSERT into wrhwr  VALUES (3,'MI-EXT-0001',date'2013-06-15','Com','VER-EXC','Revisione')
    INSERT into wrhwr  VALUES (4,'MI-EXT-0001',date'2013-06-25','Com','VER-EXC','Verifica')
    INSERT into wrhwr  VALUES (5,'MI-EXT-0001',date'2013-04-14','Com','VER-EXC','Revisione')
    INSERT into wrhwr  VALUES (6,'MI-EXT-0001',date'2013-04-30','Com','VER-EXC','Verifica')
    INSERT into wrhwr  VALUES (7,'MI-EXT-0001',date'2013-03-14','Com','VER-EXC','Collaudo')
    Query used
    SELECT *
      FROM (SELECT eq_id,
                   date_completed,
                   RANK ()
                   OVER (PARTITION BY eq_id
                         ORDER BY date_completed DESC NULLS LAST)
                      rn
              FROM wrhwr
             WHERE     status != 'S'
                   AND pmp_id LIKE 'VER-EX%'
                   AND description LIKE '%Verifica%') table1,
           (SELECT eq_id,
                   date_completed,      
                   RANK ()
                   OVER (PARTITION BY eq_id
                         ORDER BY date_completed DESC NULLS LAST)
                      rn
              FROM wrhwr
             WHERE     status != 'S'
                   AND pmp_id LIKE 'VER-EX%'
                   AND description LIKE '%Revisione%') table2,
           (SELECT eq_id,
                   date_completed,           
                   RANK ()
                   OVER (PARTITION BY eq_id
                         ORDER BY date_completed DESC NULLS LAST)
                      rn
              FROM wrhwr
             WHERE     status != 'S'
                   AND pmp_id LIKE 'VER-EX%'
                   AND description LIKE '%Collaudo%') table3
    WHERE     table1.eq_id = table3.eq_id
           AND table2.eq_id = table3.eq_id
           AND table1.eq_id = table2.eq_id
    Purpose of the above query is to selfjoin wrhwr table 3 times in order to have for every row:
    eq_id;
    completition date of a work request of type Verifica for this eq_id (table1 alias);
    completition date of a wr of type Revisione (table2 alias) for this eq_id;
    completition date of a wr of type Collaudo (table3 alias) for this eq_id;
    A distinct eq_id:
    can have many work requests (wrhwr records) with different completition dates or without completition date (date_completed column NULL);
    in a date range can have all the types of wrhwr ('Verifica', 'Revisione', 'Collaudo') or some of them (ex. Verifica, Revisione but not Collaudo, Collaudo but not Verifica and Revisione, etc.);
    substrings in description shouldn't repeat;
    (eq_id,date_completed) aren't unique but (eq_id,date_completed,description,pmp_id) should be unique;
    Expected output
    Using sample data above I expect this output:
    eq_id,table1.date_completed,table2.date_completed,table3.date_completed
    MI-EXT-001,2013-07-01,2013-07-16,2013-03-14 <--- for this eq_id table3 doesn't have 3 rows but only 1. I want to repeat the most ranked value of table3 for every result row
    MI-EXT-001,2013-07-01,2013-06-15,2013-03-14 <-- I don't wanna this row because table1 and table2 have both 3 rows so the match must be in rank terms (1st,1st) (2nd,2nd) (3rd,3rd)
    MI-EXT-001,2013-06-25,2013-06-15,2013-03-14 <-- 2nd rank of table1 joins 2nd rank of table2
    MI-EXT-001,2013-04-30,2013-04-14,2013-03-14 <-- 1st rank table1, 1st rank table2, 1st rank table3
    In vector style syntax, expected tuple output must be:
    ix = i-th ranking of tableX
    (i1, i2, i3) IF EXISTS an i-th ranking row in every table
    ELSE
    (i1, b, b)
    where b is the first available lower ranking of table2 OR NULL if there isn't any row  of lower ranking.
    Any clues?
    With the query I'm unable to remove "spurius" rows.
    I'm thinking at a solution based on analytic functions like LAG() and LEAD(), using ROLLUP() or CUBE(), using nested query but I would find a solution elegant, easy, fast and easy to maintain.
    Thanks

    FrankKulash ha scritto:
    About duplicate dates: I was most interested in what you wanted when 2 (or more) rows with the same eq_id and row type (e.g. 'Collaudo') had exactly the same completed_date.
    In the new results, did you get the columns mixed up?  It looks like the row with eq_id='MI-EXT-0002' has 'Collaudo' in the desciption, but the date appears in the verifica column of the output, not the collaudo  column.
    Why don't you want 'MI-EXT-0001' in the results?  Is it realted to the non-unique date?
    For all optimization questions, see the forum FAQ:https://forums.oracle.com/message/9362003
    If you can explain what you need to do in the view (and post some sample data and output as examples) then someone might help you find a better way to do it.
    It looks like there's a lot of repetition in the code.  Whatever you're trying to do, I suspect there's a simpler, more efficient way to do it.
    About Duplicate dates: query must show ONLY one date_completed and ignore duplicated. Those records are "bad data". You can't have 2 collaudos with the same date completed.
    Collaudo stands for equipment check. A craftperson does an equipment check once a day and, with a mobile app, update the work request related to equipment and procedure of preventive maintenance, so is impossibile that complete more than one check (Collaudo) in a day, by design.
    In the new results, it's my fault: during digitation I've swapped columns
    With "I don't want 'MI-EXT-0001'" I mean: "I don't want to show AGAIN MI-EXT-0001. In the previous post was correct the output including MI-EXT-0001.
    Regarding optimisation...
    repetition of
    LAST_VALUE ( 
                            MIN (CASE WHEN r_type = THEN column_name END) IGNORE NULLS) 
                         OVER (PARTITION BY eq_id ORDER BY r_num)  AS alias_column_name
    is because I don't know another feasible way to have all columns needed of table wrhwr in main query, maintaining the correct order. So i get them in got_r_type and propagate them in all the subquery.
    In main query I join eq table (which contains all the information about a specific equipment) with "correct" dates and columns of wrhwr table.
    I filter eq table for the specific equipment standard (eq_std column).
    efm_eq_tablet table and where clause
    AND e.eq_id = e2.eq_id 
              AND e2.is_active = 'S'; 
    means: show only rows in eq table that have an equal row in efm_eq_tablet table AND are active (represented by 'S' value in is_active column).
    About the tables v2, r2 and c2
              (SELECT doc_data, doc_data_rinnovo, eq_id 
                 FROM efm_doc_rep edr 
                WHERE edr.csi_id = '1011503' AND edr.doc_validita_temp = 'LIM') v2, 
              (SELECT doc_data, doc_data_rinnovo, eq_id 
                 FROM efm_doc_rep edr 
                WHERE     eq_id = edr.eq_id 
                      AND edr.csi_id = '1011504' 
                      AND edr.doc_validita_temp = 'LIM') r2, 
              (SELECT doc_data, doc_data_rinnovo, eq_id 
                 FROM efm_doc_rep edr 
                WHERE     edr.csi_id IN ('1011505', '1011507') 
                      AND edr.doc_validita_temp = 'LIM' 
                      AND adempimento_ok = 'SI') c2, 
    Those tables contains "alternate" dates of completition to be used when there isn't any wrhwr row for an eq_id OR when all date_completed are NULL.
    NVL() and NVL2() functions are used in main query in order to impletement this.
    The CASE/WHEN blocks inside main query implements the behavior of selecting the correct date based of the above conditions.

  • How i can pay again creative cloud with the same number account.

    Good afternoon, i would like to know how i can pay again creative cloud if the last time they try to get the money from my bank was imposible cause was empty. Now i have the money in the bank with the same account, but its automatic so now i cant do it for myself. Somebody can help me? thanks so much. Cesar Cordero

    Hi Cesar,
    Please contact the support through call/chat for billing issues.
    Contact Customer Care
    Regards,
    Sheena

Maybe you are looking for

  • Cups filter problems?

    upgrade to cups has my network printing broken: logs Leave this: [Job 292] Running /usr/bin/gs -dQUIET -dPARANOIDSAFER -dNOPAUSE -dBATCH -sDEVICE=pxlcolor -sstdout=%stderr -r600 -dDEVICEWIDTHPOINTS=612 -dDEVICEHEIGHTPOINTS=792 -dMediaPosition=0 -dBit

  • Script to insert new master on every other page

    I've got an indesign document with 131 pages (front side of postcards). I need to add the back side of the postcard (Master) to every other page. I don't want to manually do this due to the time and that I will need to do this again and again. Does a

  • Nokia 6124 classic wont work with the USB Data Cab...

    Hi all, I'm trying to connett Nokia 6124 classic but its not working with the USB Data Cable. Is there a way to fixing it yourself? Thanks Jonathan 'Jonnie'

  • Bootcamp or Parallels for windows games?

    I only use Bootcamp generally to play windows games on my Mac. It is a pain to have to constantly reboot each time to play the game and then switch back to my Mac side and vice versa. I just learned about Parallels for mac. Is it worth it to change t

  • Generic Data source Issue with Infoset

    Hi Experts, I have created generic data source with Infoset.As per customer requirement, I have to modify join condition between tables from equal to left outer join in infoset.I have tested scenario successfully using infoset query. I have done belo