GTA:V Character transfer

Does anybody know whether it is possible to transfer a GTA V character from an Xbox One console onto a PS4 console? I know that it was possible to go from Xbox 360 to PS4 (or PS3 to Xbox One) but appreciate that this transfer was from previous Gen to current Gen consoles...so therefore don't know if the cross platform transfer would still be possible between two current generation consoles.
If anybody has had any experience with this, or knows of anybody that has done (or not being able to do) this in the past then that would be greatly appreciated!
Thanks!

Thanks, got the answer here for anybody also needing to know:
"Unfortunately, you cannot transfer GTA Online progress between the Xbox One and PS4."
https://support.rockstargames.com/hc/communities/public/questions/204601177-GTA-V-Character-transfer

Similar Messages

  • Some character cannot  be transfered by Transformer

    as i transfer a document to string, the character transfer to "𦯳"
    , the original word is "��"
    how can i solve the problems,
    thanks

    thanks for your reply,
    the charset was made by others,
    the problem i find is that,
    eg. the word, "��" , after transfer, it become "&1258;" form as i write in a text file
    (this example is not real, since the character cannot display here),
    but as i use brower to view , it works properly.

  • I as paid but not received gta cash. Shark whale Card

    I purchased a Whale Shark Cash Card the previous night worth €37,99 through Playstation store [not from GTA Online gameplay. Seperately from the store]. But I still didn't get the money on my online charector....it has been more than 72 hours since I purchased it. When I purchased it, it showed 'Thank you for purchasing' etc. Then I went to my GTA Online character, the money's not there since.

    TerryTibbs85 wrote:
    Same thing happened to me just now £31.99 spent and not a dollar in my characters account!! What the F**k is going on Playstation!!!! How do you get this resolved???Check the product code on your game disk and game case spine a UK PS3 copy of GTA will be BLES 01807,if the code is different the Shark card won't work because you have an imported copy that doesn't match your countries PSN store. Not sure on product codes on PS4 think they're slightly different,but on PS3 BLES =  EU,BLAS = Asia and BLUS = US,with the BLES European codes the numbers afterwards can vary from country to country. Although PlayStation games are region free additional content never is 99% of the time so always check you have a local copy and avoid importing if you want to purchase additional content.

  • Error while running service tax program (New)

    Hi
    I have created S3 as interim tax code and S4 as final tax code. I assigned S4 to S3 as a target tax code. Entries are perfectly flowing.
    But when i do service tax program for, session is created properly. When i run it in SM35, for gived error
    "Tax code S4 may only contain one assignment line"
    I have given routine 166 in OBQ3 for ECess and SHEC.
    I tells me to do following procedure:
    Use a different tax code.
    For direct postings to a tax account, you might have to define a separate tax code for which only one line with a percentage rate other than zero is active in percentage rate maintenance. All other lines must be inactive.
    Please help me.
    Edited by: Deepak Agrawal on May 4, 2010 1:29 PM

    There is one interim service tax code (for normal services) - S1. Final service tax code S2 is created assigned to S1 as target tax code.
    Entry passed:
    Exp A/c.. Dr                                   500000
    Service Tax Basic (deferred)       50000
    Service Tax ECS (deferred)         1000
    Service Tax SHEC (deferred)       500
       To Vendor A/c                                        551500
    For this service tax program (new) is working fine.
    For the GTA scenario where service tax is not charged in invoice but we have to pay to government, in that case i have created one interim tax code G1 and i have assigned final tax code S2.
    Entry passed for GTA (invoice raised by transporter is Rs. 500000):
    Exp A/c.. Dr                                   515000
       To Service Tax Basic (deferred)       50000
       To Service Tax ECS (deferred)         1000
       To Service Tax SHEC (deferred)       500
       To Vendor A/c                                   500000
    But when i execute the program i can see three line items (for basic, interim and SHEC)
    But when i run the session, it gives error:
    Tax code S2 may only contain one assignment line.
    I AM UNABLE TO UNDERSTAND, WHY ITS GIVING ERROR FOR GTA SERVICE TAX TRANSFER
    Edited by: Meenu_ND on Jun 1, 2010 10:48 AM

  • Transfer a caller directly to a Unity Express Mailbox using a prepend * character

    I have a CME and CUE setup for our phone system. I experiemented setting up the feature to automatically transfer a calle directly to a CUE mailbox.
    In CME
    ephone-dn 128
    number 32252
    call-forward all 2999
    In CUE
    At config prompt
    username JoeWho phonenumberE164 32252
    This set up worked and calls/transfers using 32252 went directly to the voicemail. However, is it possible to use * as the prepend and not a number. A comment in the following article suggests that you can:
    http://www.tech-recipes.com/rx/2287/callmanager_express_transfer_direct_to_unity_express_voicemail/
    However, when I tried using the *, CUE objected saying that * was an invalid character.
    Here is what I am running:
    VicCME#sh telephony-service
    CONFIG (Version=7.1)
    =====================
    Version 7.1
    Cisco Unified Communications Manager Express
    VicCUE# sh software version
    Cisco Unity Express version (7.0.3)
    Technical Support: http://www.cisco.com/techsupport Copyright (c) 1986-2008 by Cisco Systems, Inc.
    Components:
    - CUE Voicemail Language Support version  7.0.3

    Thank you, Rob
    Yes, I was actually doing more work than required. I cloned the production template and created a new template#3 and added the trnsfvm option as follows to the softkeys line of the template. I then applied that to my phone and reset my phone as shown below (Setting up softkey for trnsfvm). The transfer to voicemail works perfectly using the softkey. However, I cannot seem to get the fac option to work. That is, suppose I want to leave a message for someone in their voicemail (not transfering an outside call to someone's voicemail). For example, I have tried many keys my config is:
    telephony-service
    fac custom trnsfvm * (#* or #0 or #22)
    I reset my phone and when I press the fac keys, I receive an undefined # on my phone. For example, if I define the fac as #0, then I would pick up the headset hit the #0extension# keys. So, if I wanted to leave a message to extension 2148's voicemail, I would dial #02148. However, this does not work.
    Was there something else that I needed to do? Do I have to create the cnf-files?
    telephony-service
    create cnf-files
    Or, do I have to reset the CME/phone system?
    I also tried "fac-standard" that sets trnsfvm to *6, but that also does not work and I also get an "Undefined" display when I try to use it.
    VicCME#sh telephony-service fac
      telephony-service fac standard
        callfwd all **1
        callfwd cancel **2
        pickup local **3
        pickup group **4
        pickup direct **5
        park **6
        dnd **7
        redial **8
        voicemail **9
        ephone-hunt join *3
        ephone-hunt cancel #3
        ephone-hunt hlog *4
        ephone-hunt hlog-phone *5
        trnsfvm *6
        dpark-retrieval *0
    VicCME#
    Setting up the softkey for Trnsfvm.
    phone-template  3
    conference drop-mode local
    conference add-mode creator
    conference admin
    softkeys hold  Resume Newcall
    softkeys idle  Redial Newcall Cfwdall Dnd
    softkeys seized  Redial Endcall Meetme
    softkeys alerting  Endcall
    softkeys connected  Hold Trnsfer Endcall Trnsfvm LiveRcd Confrn ConfList
    softkeys ringing  Answer Dnd
    ephone  28
    ephone-template 3

  • Codename "Data Transfer" Won't Accept # Character in Server Password

    Administrator password requirements for SQL Azure server instances require any three of lowercase letters, uppercase letters, numbers and/or symbols. I ordinarily use a # symbol in my database admin passwords.
    I'm unable to log in to my database to upload *.csv data because attempts to connect with the pass#1234 style fail. Fiddler2 shows the password is truncated at pass. (This is an example, not  my password.)
    This needs fixing, as does forcing me to create a new profile name when I log in to the forum.
    --rj

    Hey there, and apologies.  I wanted to reach out and see if this is still a problem for you.
    Thanks Guy

  • Issues with using the output redirection character with newer NXOS versions?

    Has anyone seen any issues with using the output redirection character with newer NXOS versions?
    Am receiving "Error 0x40870004 while copying."
    Simply copying a file from bootflash to tftp is ok.
    This occurs for both 3CDaemon and Tftpd32 softwares.
    Have tried it on multiple switches - same issue.
    Any known bugs?
    thanks!
    The following is an example of bad (NXOS4.1.1b) and good (SANOS3.2.1a)
    MDS2# sho ver | inc system
      system:    version 4.1(1b)
      system image file is:    bootflash:///m9200-s2ek9-mz.4.1.1b.bin
      system compile time:     10/7/2008 13:00:00 [10/11/2008 09:52:55]
    MDS2# sh int br > tftp://10.73.54.194
    Trying to connect to tftp server......
    Connection to server Established. Copying Started.....
    TFTP put operation failed:Access violation
    Error 0x40870004 while copying tftp://10.73.54.194/
    MDS2# copy bootflash:cpu_logfile tftp://10.73.54.194
    Trying to connect to tftp server......
    Connection to server Established. Copying Started.....
    |
    TFTP put operation was successful
    MDS2#
    ck-ci9216-001# sho ver | inc system
      system:    version 3.2(1a)
      system image file is:    bootflash:/m9200-ek9-mz.3.2.1a.bin
      system compile time:     9/25/2007 18:00:00 [10/06/2007 06:46:51]
    ck-ci9216-001# sh int br > tftp://10.73.54.194
    Trying to connect to tftp server......
    |
    TFTP put operation was successful

    Please check with new version of TFTPD 32 server. The error may be due to older version of TFPT server, the new version available solved this error. Files are getting uploaded with no issues.
    1. Download tftpd32b.zip from:
    http://tftpd32.jounin.net/tftpd32_download.html
    2. Copy the tftpd32b.zip file into an empty directory and extract it.
    3. Copy the file you want to transver into the directory containing tftpd32.exe.
    4. Run tftpd32.exe from that directory. The "Base Directory" field should show the path to the directory containing the file you want to transfer.
    At this point, the tftpserver is ready to begin serving files. As devices request files, the main tftpd32 window will log the requests.
    Best Regards...

  • I can't figure out how to transfer my albums on my iphone to photo stream so that I can see them on my apple computer. It did not automatically sync.

    I can't figure out how to transfer my albums on my iphone to photo stream so that I can see them on my apple computer. It did not automatically sync.

    kellyfromdanville wrote:
    I am new to Apple!
    Ah, new to routers too.
    Perhaps it's best to have a local computer support person come set you up?
    To explain everything would require writing a book here.
    But this is what you want basically.
    Router, new Wireless N, firmware updated.
    1: Two accounts, one admin, one guest Internet access. Random +30 plus character lenght passwords each preferred.
    (guest password to max password length of any iDevices)
    2: Encryption: WPA2 AES. No Ping. No remote access. No MAC filtering. Visible.
    3: Guest access on devices and computers only.  Passwords written down and stored.
    4: OpenDNS preferred but not mandatory.
    If you get OpenDNS, you can log into their website for free and set parental controls on things like inappropriate sites for children. So whatever device is used on your network can't access a large portion of those sites (nothing is perfect) by most users who are not technical oriented.
    What your doing now is running the bare bones unsecure router which can be used harmfully against you.
    Even a Apple Airport router has to be very well secured. The defautl cd and out of the box setup is not a secure solution.

  • Phone numbers with "pause character" not working same as first-gen iPhone

    Is this a bug?
    When I configure a phone number in my contact list to use the "pause" character (which shows up as a comma character in the display) a number of strange things happen.
    For instance, if I configure the following: (###) ###-#### ,,**2216464#
    where (###) ###-#### is my office phone number
    ,, is a pause to allow for the automated attendent to pick-up
    ** is recognized by my phone system as a transfer to my voice mailbox
    2216464 is my voice mailbox
    # is a termination character recognized by my phone system.
    Using my first-generation iPhone, this worked fine.
    Using my new iPhone 3G, this no longer works.
    Some observations:
    When directly using the touchpad on the iPhone, it appears to send DTMF tones as quickly and for as long as you hold your finger on the keypad. But, if you configure a phone number in the contact list and that phone number has a pause in it, each number configured after the pause will actuate for several seconds. The tone for each number (in this case, the 2216464) last SO long that my phone system gives up on the call and, basically, terminates the call. In addition, the inter-digit time is SO small as to be non-existant. In other words, in my sample phone number, the leading two digit of my mailbox number sound like they are expressed by the iPhone as a single, extra-long tone. There should be a longer pause between the first two digits. If I attempt to break them apart by inserting a pause (displayed as a comma), the pause is SO long that, again, my phone system gives up.
    So, in summary:
    Why did this change between the first generation and second generation iPhones?
    Also, why is the pause "character" SO LONG? Instead of making a pause character, say, four seconds, why not make it one second and let the user put in four pauses if they want a four-second pause?
    Why does the character length appear to extend after use of a pause character?
    Why is the inter-digit pause practically non-existant?
    Thanks.
    Ken

    I'm having the same issue. I just got the phone last week and was trying to call a number with an extension. After realizing that Address book does not have an extension field I reformatted the phone number using the "pause" (comma). With this in place it will not dial the number. Remove the pause however, and it will dial the root phone number.
    Please tell me this is just a bug and will be fixed soon. It's bad enough that I've had to spend the last week reworking my contacts in Address Book after having to jump ship from my Palm software. As a business user it is a regular piece of daily business to encounter the automated receptionist asking for someone's extension. Since the iPhone doesn't show the number you just dialed (the extension is in there after the "x" but it doesn't recognize it) you end up having to memorize the extension prior to making each call.

  • How to delete blank space from value of Infoobject in the transfer rule

    does anyboady know how to write a routine in the transfer rule in order to delete ending blank spaces from value of Infoobject 'ORDTEXT'. E.g. value: 'ABM    '?  After routine, the result should be 'ABM'.
    Thanks for any input.

    Deleting 'ending blank spaces' from a character type field (ie info-object or text) in SAP is not meaningful. For all usage, a char type field of length 8 having value  'ABM     '  is same as 'ABM'.
    Ok, HTML is changing the display in the above, so let me restate it
    For all usage, a char type field of length 8 having value  'ABM_____'  is same as 'ABM', take the underscore character (_) in the first string as a placeholder for space character.
    Edited by: Ajay Das on Jul 29, 2009 12:07 PM

  • Invalid character error while activating DSO - Please help

    Hi experts,
    I'm getting invalid hex character error when loading a text infoobject in my DSO. Repairing data records in the PSA is not an option. Changing the data in the source system is also not an option. I've tried many of the methods suggested in other threads, but am unable to get it working (ABAP code, "ALL_CAPITAL" etc). We are on an unicode system.
    Can someone provide some clean code to eliminate the illegal hex characters (00-1f) in my transfer rule routine. Basically, replace the illegal hex characters with space or an acceptable character.
    Many thanks,
    Anita S.

    Hi,
    I have used the below code for removing all unwanted chars from 0MATERIAL, that worked fine. You can modify it for ur need.
      CONSTANTS:
           G_ALLOWED_CHAR(58) TYPE C VALUE
             ' !"%&''()*+,-./:;<=>?_0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZ'.
        data : E_ALLOWED_CHAR(58).
    RESULT = SOURCE_FIELDS-MATNR.
        SHIFT RESULT RIGHT DELETING TRAILING SPACE.
        SHIFT RESULT LEFT DELETING LEADING SPACE.
        TRANSLATE RESULT TO upper case.
    *RESULT = comm_structure-/BIC/ZCONTRTD.
        while RESULT ca '#'.
          RESULT+sy-fdpos(1) = ' '.
        endwhile.
        if RESULT(1) = '!'.
          RESULT(1) = ' '.
        endif.
        do.
          if not RESULT co g_allowed_char.
            shift RESULT+sy-fdpos left.
          else.
            exit.
          endif.
        enddo.
    Thanks
    dipika

  • ORA-12712: new character set must be a superset of old character ...-

    Dear all ,
    i have a problem with oracle database,
    server configuration oracle 10g, solaris 9, sparc processor 64 bit,
    i altered database character set to UTF8, whe i return database to old seeting ISO 8859-6
    i got this error ORA-12712: new character set must be a superset of old character set.
    anyone can help me and let me know how can i solve this error?
    thanks
    Gasir

    You should recreate a new database and use export/import to transfer data from the existing UTF8 database to the new ISO 8859-6 database according to the ]Globalization Guide.

  • Error in transfer structure while activating the business content

    Hi all
            we got the following error while activating the transfer rules( business content) :
    1)error when determining a number from object BI_TSDTEL and number 01
    2) object name can only contain characters from syntactical character set.
    3) data element for infoobject WAERS 0CURRENCY could not be created.
    can anyone please explain what is this error.

    Hello Tilak,  
    Goto SE37 -> RSAR_TRANSTRUCTURE_ACTIVATE -> Transfer Structure Name -> Execute
    If the above doesn't fix then try the below
    Goto SE14 -> Transfer Structure Name -> Adjust and activate .
    This will fix if there is any inconsistency in the transfer structure.
    [Thanks|http://chandranonline.blogspot.com/]
    [Chandran|http://chandranonline.blogspot.com/]

  • Fixing a US7ASCII - WE8ISO8859P1 Character Set Conversion Disaster

    In hopes that it might be helpful in the future, here's the procedure I followed to fix  a disastrous unintentional US7ASCII on 9i to WE8ISO8859P1 on 10g migration.
    BACKGROUND
    Oracle has multiple character sets, ranging from US7ASCII to AL32UTF16.
    US7ASCII, of course, is a cheerful 7 bit character set, holding the basic ASCII characters sufficient for the English language.
    However, it also has a handy feature: character fields under US7ASCII will accept characters with values > 128. If you have a web application, users can type (or paste) Us with umlauts, As with macrons, and quite a few other funny-looking characters.
    These will be inserted into the database, and then -- if appropriately supported -- can be selected and displayed by your app.
    The problem is that while these characters can be present in a VARCHAR2 or CLOB column, they are not actually legal. If you try within Oracle to convert from US7ASCII to WE8ISO8859P1 or any other character set, Oracle recognizes that these characters with values greater than 127 are not valid, and will replace them with a default "unknown" character. In the case of a change from US7ASCII to WE8ISO8859P1, it will change them to 191, the upside down question mark.
    Oracle has a native utility, introduced in 8i, called csscan, which assists in migrating to different character sets. This has been replaced in newer versions with the Database MIgration Assistant for Unicode (DMU), which is the new recommended tool for 11.2.0.3+.
    These tools, however, do no good unless they are run. For my particular client, the operations team took a database running 9i and upgraded it to 10g, and as part of that process the character set was changed from US7ASCII to WE8ISO8859P1. The database had a large number of special characters inserted into it, and all of these abruptly turned into upside-down question marks. The users of the application didn't realize there was a problem until several weeks later, by which time they had put a lot of new data into the system. Rollback was not possible.
    FIXING THE PROBLEM
    How fixable this problem is and the acceptable methods which can be used depend on the application running on top of the database. Fortunately, the client app was amenable.
    (As an aside note: this approach does not use csscan -- I had done something similar previously on a very old system and decided it would take less time in this situation to revamp my old procedures and not bring a new utility into the mix.)
    We will need to separate approaches -- one to fix the VARCHAR2 & CHAR fields,  and a second for CLOBs.
    In order to set things up, we created two environments. The first was a clone of production as it is now, and the second a clone from before the upgrade & character set change. We will call these environments PRODCLONE and RESTORECLONE.
    Next, we created a database link, OLD6. This allows PRODCLONE to directly access RESTORECLONE. Since they were cloned with the same SID, establishing the link needed the global_names parameter set to false.
    alter system set global_names=false scope=memory;
    CREATE PUBLIC DATABASE LINK OLD6
    CONNECT TO DBUSERNAME
    IDENTIFIED BY dbuserpass
    USING 'restoreclone:1521/MYSID';
    Testing the link...
    SQL> select count(1) from users@old6;
      COUNT(1)
           454
    Here is a row in a table which contains illegal characters. We are accessing RESTORECLONE from PRODCLONE via our link.
    PRODCLONE> select dump(title) from my_contents@old6 where pk1=117286;
    DUMP(TITLE)
    Typ=1 Len=49: 78,67,76,69,88,45,80,78,174,32,69,120,97,109,32,83,116,121,108,101
    ,32,73,110,116,101,114,97,99,116,105,118,101,32,82,101,118,105,101,119,32,81,117
    ,101,115,116,105,111,110,115
    By comparison, a dump of that row on PRODCLONE's my_contents gives:
    PRODCLONE> select dump(title) from my_contents where pk1=117286;
    DUMP(TITLE)
    Typ=1 Len=49: 78,67,76,69,88,45,80,78,191,32,69,120,97,109,32,83,116,121,108,101
    ,32,73,110,116,101,114,97,99,116,105,118,101,32,82,101,118,105,101,119,32,81,117
    ,101,115,116,105,111,110,115
    Note that the "174" on RESTORECLONE was changed to "191" on PRODCLONE.
    We can manually insert CHR(174) into our PRODCLONE and have it display successfully in the application.
    However, I tried a number of methods to copy the data from RESTORECLONE to PRODCLONE through the link, but entirely without success. Oracle would recognize the character as invalid and silently transform it.
    Eventually, I located a clever workaround at this link:
    https://kr.forums.oracle.com/forums/thread.jspa?threadID=231927
    It works like this:
    On RESTORECLONE you create a view, vv, with UTL_RAW:
    RESTORECLONE> create or replace view vv as select pk1,utl_raw.cast_to_raw(title) as title from my_contents;
    View created.
    This turns the title to raw on the RESTORECLONE.
    You can now convert from RAW to VARCHAR2 on the PRODCLONE database:
    PRODCLONE> select dump(utl_raw.cast_to_varchar2 (title)) from vv@old6 where pk1=117286;
    DUMP(UTL_RAW.CAST_TO_VARCHAR2(TITLE))
    Typ=1 Len=49: 78,67,76,69,88,45,80,78,174,32,69,120,97,109,32,83,116,121,108,101
    ,32,73,110,116,101,114,97,99,116,105,118,101,32,82,101,118,105,101,119,32,81,117
    ,101,115,116,105,111,110,115
    The above works because oracle on PRODCLONE never knew that our TITLE string on RESTORE was originally in  US7ASCII, so it was unable to do its transparent character set conversion.
    PRODCLONE> update my_contents set title=( select utl_raw.cast_to_varchar2 (title) from vv@old6 where pk1=117286) where pk1=117286;
    PRODCLONE> select dump(title) from my_contents where pk1=117286;
    DUMP(UTL_RAW.CAST_TO_VARCHAR2(TITLE))
    Typ=1 Len=49: 78,67,76,69,88,45,80,78,174,32,69,120,97,109,32,83,116,121,108,101
    ,32,73,110,116,101,114,97,99,116,105,118,101,32,82,101,118,105,101,119,32,81,117
    ,101,115,116,105,111,110,115
    Excellent! The "174" character has survived the transfer and is now in place on PRODCLONE.
    Now that we have a method to move the data over, we have to identify which columns /tables have character data that was damaged by the conversion. We decided we could ignore anything with a length smaller than 10 -- such fields in our application would be unlikely to have data with invalid characters.
    RESTORECLONE> select count(1) from user_tab_columns where data_type in ('CHAR','VARCHAR2') and data_length > 10;
       COUNT(1)
        533
    By converting a field to WE8ISO8859P1, and then comparing it with the original, we can see if the characters change:
    RESTORECLONE> select count(1) from my_contents where title != convert (title,'WE8ISO8859P1','US7ASCII') ;
      COUNT(1)
         10568
    So 10568 rows have characters which were transformed  into 191s as part of the original conversion.
    [ As an aside, we can't use CONVERT() on LOBs -- for them we will need another approach, outlined further below.
    RESTOREDB> select count(1) from my_contents where main_data != convert (convert(main_DATA,'WE8ISO8859P1','US7ASCII'),'US7ASCII','WE8ISO8859P1') ;
    select count(1) from my_contents where main_data != convert (convert(main_DATA,'WE8ISO8859P1','US7ASCII'),'US7ASCII','WE8ISO8859P1')
    ERROR at line 1:
    ORA-00932: inconsistent datatypes: expected - got CLOB
    Anyway, now that we can identify VARCHAR2 fields which need to be checked, we can put together a PL/SQL stored procedure to do it for us:
    create or replace procedure find_us7_strings
    (table_name varchar2,
    fix_col varchar2 )
    authid current_user
    as
    orig_sql varchar2(1000);
    begin
    orig_sql:='insert into cnv_us7(mytablename,myindx,mycolumnname)  select '''||table_name||''',pk1,'''||fix_col||''' from '||table_name||' where '||fix_col||' !=  CONVERT(CONVERT('||fix_col||',''WE8ISO8859P1''),''US7ASCII'') and '||fix_col||' is not null';
    -- Uncomment if debugging:
    -- dbms_output.put_line(orig_sql);
      execute immediate orig_sql;
    end;
    And create a table to store the information as to which tables, columns, and rows have the bad characters:
    drop table cnv_us7;
    create table cnv_us7 (mytablename varchar2(50), myindx number,      mycolumnname varchar2(50) ) tablespace myuser_data;
    create index list_tablename_idx on cnv_us7(mytablename) tablespace myuser_indx;
    With a SQL-generating SQL script, we can iterate through all the tables/columns we want to check:
    --example of using the data: select title from my_contents where pk1 in (select myindx from cnv_us7)
    set head off pagesize 1000 linesize 120
    spool runme.sql
    select 'exec find_us7_strings ('''||table_name||''','''||column_name||'''); ' from user_tab_columns
          where
              data_type in ('CHAR','VARCHAR2')
              and table_name in (select table_name from user_tab_columns where column_name='PK1' and  table_name not  in ('HUGETABLEIWANTTOEXCLUDE','ANOTHERTABLE'))
              and char_length > 10
              order by table_name,column_name;
    spool off;
    set echo on time on timing on feedb on serveroutput on;
    spool output_of_runme
    @./runme.sql
    spool off;
    Which eventually gives us the following inserted into CNV_US7:
    20:48:21 SQL> select count(1),mycolumnname,mytablename from cnv_us7 group by mytablename,mycolumnname;
             4 DESCRIPTION                                        MY_FORUMS
         21136 TITLE                                              MY_CONTENTS
    Out of 533 VARCHAR2s and CHARs, we only had five or six columns that needed fixing
    We create our views on  RESTOREDB:
    create or replace view my_forums_vv as select pk1,utl_raw.cast_to_raw(description) as description from forum_main;
    create or replace view my_contents_vv as select pk1,utl_raw.cast_to_raw(title) as title from my_contents;
    And then we can fix it directly via sql:
    update my_contents taborig1 set TITLE= (select utl_raw.cast_to_varchar2 (TITLE) from my_contents_vv@old6 where pk1=taborig1.pk1)
    where pk1 in (
    select tabnew.pk1 from my_contents@old6 taborig,my_contents tabnew,cnv_us7@old6
          where taborig.pk1=tabnew.pk1
              and myindx=tabnew.pk1
              and mycolumnname='TITLE'
              and mytablename='MY_CONTENTS'
              and convert(taborig.TITLE,'US7ASCII','WE8ISO8859P1') = tabnew.TITLE );
    Note this part:
          "and convert(taborig.TITLE,'US7ASCII','WE8ISO8859P1') = tabnew.TITLE "
    This checks to verify that the TITLE field on the PRODCLONE and RESTORECLONE are the same (barring character set issues). This is there  because if the users have changed TITLE  -- or any other field -- on their own between the time of the upgrade and now, we do not want to overwrite their changes. We make the assumption that as part of the process, they may have changed the bad character on their own.
    We can also create a stored procedure which will execute the SQL for us:
    create or replace procedure fix_us7_strings
    (TABLE_NAME varchar2,
    FIX_COL varchar2 )
    authid current_user
    as
    orig_sql varchar2(1000);
    TYPE cv_type IS REF CURSOR;
    orig_cur cv_type;
    begin
    orig_sql:='update '||TABLE_NAME||' taborig1 set '||FIX_COL||'= (select utl_raw.cast_to_varchar2 ('||FIX_COL||') from '||TABLE_NAME||'_vv@old6 where pk1=taborig1.pk1)
    where pk1 in (
    select tabnew.pk1 from '||TABLE_NAME||'@old6 taborig,'||TABLE_NAME||' tabnew,cnv_us7@old6
          where taborig.pk1=tabnew.pk1
              and myindx=tabnew.pk1
              and mycolumnname='''||FIX_COL||'''
              and mytablename='''||TABLE_NAME||'''
              and convert(taborig.'||FIX_COL||',''US7ASCII'',''WE8ISO8859P1'') = tabnew.'||FIX_COL||')';
    dbms_output.put_line(orig_sql);
    execute immediate orig_sql;
    end;
    exec fix_us7_strings('MY_FORUMS','DESCRIPTION');
    exec fix_us7_strings('MY_CONTENTS','TITLE');
    commit;
    To validate this before and after, we can run something like:
    select dump(description) from my_forums where pk1 in (select myindx from cnv_us7@old6 where mytablename='MY_FORUMS');
    The above process fixes all the VARCHAR2s and CHARs. Now what about the CLOB columns?
    Note that we're going to have some extra difficulty here, not just because we are dealing with CLOBs, but because we are working with CLOBs in 9i, whose functions have less CLOB-related functionality.
    This procedure finds invalid US7ASCII strings inside a CLOB in 9i:
    create or replace procedure find_us7_clob
    (table_name varchar2,
    fix_col varchar2)
    authid current_user
    as
      orig_sql varchar2(1000);
      type cv_type is REF CURSOR;
      orig_table_cur cv_type;
      my_chars_read NUMBER;
      my_offset NUMBER;
      my_problem NUMBER;
      my_lob_size NUMBER;
      my_indx_var NUMBER;
      my_total_chars_read NUMBER;
      my_output_chunk VARCHAR2(4000);
      my_problem_flag NUMBER;
      my_clob CLOB;
      my_total_problems NUMBER;
      ins_sql VARCHAR2(4000);
    BEGIN
       DBMS_OUTPUT.ENABLE(1000000);
       orig_sql:='select pk1,dbms_lob.getlength('||FIX_COL||') as cloblength,'||fix_col||' from '||table_name||' where dbms_lob.getlength('||fix_col||') >0 and '||fix_col||' is not null order by pk1';
       open orig_table_cur for orig_sql;
       my_total_problems := 0;
       LOOP
            FETCH orig_table_cur INTO my_indx_var,my_lob_size,my_clob;
                    EXIT WHEN orig_table_cur%NOTFOUND;
            my_offset :=1;
            my_chars_read := 512;
            my_problem_flag :=0;
            WHILE my_offset < my_lob_size and my_problem_flag =0
                    LOOP
                    DBMS_LOB.READ(my_clob,my_chars_read,my_offset,my_output_chunk);
                    my_offset := my_offset + my_chars_read;
                    IF my_output_chunk != CONVERT(CONVERT(my_output_chunk,'WE8ISO8859P1'),'US7ASCII')
                            THEN
                            -- DBMS_OUTPUT.PUT_LINE('Problem with '||my_indx_var);
                            -- DBMS_OUTPUT.PUT_LINE(my_output_chunk);
                            my_problem_flag:=1;
                    END IF;
            END LOOP;
            IF my_problem_flag=1
                    THEN my_total_problems := my_total_problems +1;
                    ins_sql:='insert into cnv_us7(mytablename,myindx,mycolumnname) values ('''||table_name||''','||my_indx_var||','''||fix_col||''')';
                    execute immediate ins_sql;
                    END IF;
       END LOOP;
       DBMS_OUTPUT.PUT_LINE('We found '||my_total_problems||' problem rows in table '||table_name||', column '||fix_col||'.');
    END;
    And we can use SQL-generating SQL to find out which CLOBs have issues, out of all the ones in the database:
    RESTOREDB> select 'exec find_us7_clob('''||table_name||''','''||column_name||''');' from user_tab_columns where data_type='CLOB';
    exec find_us7_clob('MY_CONTENTS','DATA');
    After completion, the CNV_US7 table looked like this:
    RESTOREDB> set linesize 120 pagesize 100;
    RESTOREDB>  select count(1),mytablename,mycolumnname from cnv_us7
       where mytablename||' '||mycolumnname in (select table_name||' '||column_name from user_tab_columns
             where data_type='CLOB' )
          group by mytablename,mycolumnname;
      COUNT(1) MYTABLENAME                                        MYCOLUMNNAME
         69703 MY_CONTENTS                                  DATA
    On RESTOREDB, our 9i version, we will use this procedure (found many years ago on the internet):
    create or replace procedure CLOB2BLOB (p_clob in out nocopy clob, p_blob in out nocopy blob) is
    -- transforming CLOB to BLOB
    l_off number default 1;
    l_amt number default 4096;
    l_offWrite number default 1;
    l_amtWrite number;
    l_str varchar2(4096 char);
    begin
    loop
    dbms_lob.read ( p_clob, l_amt, l_off, l_str );
    l_amtWrite := utl_raw.length ( utl_raw.cast_to_raw( l_str) );
    dbms_lob.write( p_blob, l_amtWrite, l_offWrite,
    utl_raw.cast_to_raw( l_str ) );
    l_offWrite := l_offWrite + l_amtWrite;
    l_off := l_off + l_amt;
    l_amt := 4096;
    end loop;
    exception
    when no_data_found then
    NULL;
    end;
    We can test out the transformation of CLOBs to BLOBs with a single row like this:
    drop table my_contents_lob;
    Create table my_contents_lob (pk1 number,data blob);
    DECLARE
          v_clob CLOB;
          v_blob BLOB;
        BEGIN
          SELECT data INTO v_clob FROM my_contents WHERE pk1 = 16 ;
          INSERT INTO my_contents_lob (pk1,data) VALUES (16,empty_blob() );
          SELECT data INTO v_blob FROM my_contents_lob WHERE pk1=16 FOR UPDATE;
          clob2blob (v_clob, v_blob);
        END;
    select dbms_lob.getlength(data) from my_contents_lob;
    DBMS_LOB.GETLENGTH(DATA)
                                 329
    SQL> select utl_raw.cast_to_varchar2(data) from my_contents_lob;
    UTL_RAW.CAST_TO_VARCHAR2(DATA)
    Lorem ipsum dolor sit amet, consectetur adipisicing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam...
    Now we need to push it through a loop. Unfortunately, I had trouble making the "SELECT INTO" dynamic. Thus I used a version of the procedure for each table. It's aesthetically displeasing, but at least it worked.
    create table my_contents_lob(pk1 number,data blob);
    create index my_contents_lob_pk1 on my_contents_lob(pk1) tablespace my_user_indx;
    create or replace procedure blob_conversion_my_contents
    (table_name varchar2,
    fix_col varchar2)
    authid current_user
    as
      orig_sql varchar2(1000);
      type cv_type is REF CURSOR;
      orig_table_cur cv_type;
      my_chars_read NUMBER;
      my_offset NUMBER;
      my_problem NUMBER;
      my_lob_size NUMBER;
      my_indx_var NUMBER;
      my_total_chars_read NUMBER;
      my_output_chunk VARCHAR2(4000);
      my_problem_flag NUMBER;
      my_clob CLOB;
      my_blob BLOB;
      my_total_problems NUMBER;
      new_sql VARCHAR2(4000);
    BEGIN
      DBMS_OUTPUT.ENABLE(1000000);
       orig_sql:='select pk1,dbms_lob.getlength('||FIX_COL||') as cloblength,'||fix_col||' from '||table_name||' where pk1 in (select myindx from cnv_us7 where mytablename='''||TABLE_NAME||''' and mycolumnname='''||FIX_COL||''') order by pk1';
       open orig_table_cur for orig_sql;
       LOOP
            FETCH orig_table_cur INTO my_indx_var,my_lob_size,my_clob;
                    EXIT WHEN orig_table_cur%NOTFOUND;
            new_sql:='INSERT INTO '||table_name||'_lob(pk1,'||fix_col||') values ('||my_indx_var||',empty_blob() )';
            dbms_output.put_line(new_sql);
          execute immediate new_sql;
    -- Here's the bit that I had trouble making dynamic. Feel free to let me know what I am doing wrong.
    -- new_sql:='SELECT '||fix_col||' INTO my_blob from '||table_name||'_lob where pk1='||my_indx_var||' FOR UPDATE';
    --        dbms_output.put_line(new_sql);
            select data into my_blob from my_contents_lob where pk1=my_indx_var FOR UPDATE;
          clob2blob(my_clob,my_blob);
       END LOOP;
       CLOSE orig_table_cur;
      DBMS_OUTPUT.PUT_LINE('Completed program');
    END;
    exec blob_conversion_my_contents('MY_CONTENTS','DATA');
    Verify that things work properly:
    select dump( utl_raw.cast_to_varchar2(data))  from my_contents_lob where pk1=xxxx;
    This should let you see see characters > 150. Thus, the method works.
    We can now take this data, export it from RESTORECLONE
    exp file=a.dmp buffer=4000000 userid=system/XXXXXX tables=my_user.my_contents rows=y
    and import the data on prodclone
    imp file=a.dmp fromuser=my_user touser=my_user userid=system/XXXXXX buffer=4000000;
    For paranoia's sake, double check that it worked properly:
    select dump( utl_raw.cast_to_varchar2(data))  from my_contents_lob;
    On our 10g PRODCLONE, we'll use these stored procedures:
    CREATE OR REPLACE FUNCTION CLOB2BLOB(L_CLOB CLOB) RETURN BLOB IS
    L_BLOB BLOB;
    L_SRC_OFFSET NUMBER;
    L_DEST_OFFSET NUMBER;
    L_BLOB_CSID NUMBER := DBMS_LOB.DEFAULT_CSID;
    V_LANG_CONTEXT NUMBER := DBMS_LOB.DEFAULT_LANG_CTX;
    L_WARNING NUMBER;
    L_AMOUNT NUMBER;
    BEGIN
    DBMS_LOB.CREATETEMPORARY(L_BLOB, TRUE);
    L_SRC_OFFSET := 1;
    L_DEST_OFFSET := 1;
    L_AMOUNT := DBMS_LOB.GETLENGTH(L_CLOB);
    DBMS_LOB.CONVERTTOBLOB(L_BLOB,
    L_CLOB,
    L_AMOUNT,
    L_SRC_OFFSET,
    L_DEST_OFFSET,
    1,
    V_LANG_CONTEXT,
    L_WARNING);
    RETURN L_BLOB;
    END;
    CREATE OR REPLACE FUNCTION BLOB2CLOB(L_BLOB BLOB) RETURN CLOB IS
    L_CLOB CLOB;
    L_SRC_OFFSET NUMBER;
    L_DEST_OFFSET NUMBER;
    L_BLOB_CSID NUMBER := DBMS_LOB.DEFAULT_CSID;
    V_LANG_CONTEXT NUMBER := DBMS_LOB.DEFAULT_LANG_CTX;
    L_WARNING NUMBER;
    L_AMOUNT NUMBER;
    BEGIN
    DBMS_LOB.CREATETEMPORARY(L_CLOB, TRUE);
    L_SRC_OFFSET := 1;
    L_DEST_OFFSET := 1;
    L_AMOUNT := DBMS_LOB.GETLENGTH(L_BLOB);
    DBMS_LOB.CONVERTTOCLOB(L_CLOB,
    L_BLOB,
    L_AMOUNT,
    L_SRC_OFFSET,
    L_DEST_OFFSET,
    1,
    V_LANG_CONTEXT,
    L_WARNING);
    RETURN L_CLOB;
    END;
    And now, for the piece de' resistance, we need a BLOB to CLOB conversion that assumes that the BLOB data is stored initially in WE8ISO8859P1.
    To find correct CSID for WE8ISO8859P1, we can use this query:
    select nls_charset_id('WE8ISO8859P1') from dual;
    Gives "31"
    create or replace FUNCTION BLOB2CLOBASC(L_BLOB BLOB) RETURN CLOB IS
    L_CLOB CLOB;
    L_SRC_OFFSET NUMBER;
    L_DEST_OFFSET NUMBER;
    L_BLOB_CSID NUMBER := 31;      -- treat blob as  WE8ISO8859P1
    V_LANG_CONTEXT NUMBER := 31;   -- treat resulting clob as  WE8ISO8850P1
    L_WARNING NUMBER;
    L_AMOUNT NUMBER;
    BEGIN
    DBMS_LOB.CREATETEMPORARY(L_CLOB, TRUE);
    L_SRC_OFFSET := 1;
    L_DEST_OFFSET := 1;
    L_AMOUNT := DBMS_LOB.GETLENGTH(L_BLOB);
    DBMS_LOB.CONVERTTOCLOB(L_CLOB,
    L_BLOB,
    L_AMOUNT,
    L_SRC_OFFSET,
    L_DEST_OFFSET,
    L_BLOB_CSID,
    V_LANG_CONTEXT,
    L_WARNING);
    RETURN L_CLOB;
    END;
    select dump(dbms_lob.substr(blob2clobasc(data),4000,1)) from my_contents_lob;
    Now, we can compare these:
    select dbms_lob.compare(blob2clob(old.data),new.data) from  my_contents new,my_contents_lob old where new.pk1=old.pk1;
    DBMS_LOB.COMPARE(BLOB2CLOB(OLD.DATA),NEW.DATA)
                                                                 0
                                                                 0
                                                                 0
    Vs
    select dbms_lob.compare(blob2clobasc(old.data),new.data) from  my_contents new,my_contents_lob old where new.pk1=old.pk1;
    DBMS_LOB.COMPARE(BLOB2CLOBASC(OLD.DATA),NEW.DATA)
                                                                   -1
                                                                   -1
                                                                   -1
    update my_contents a set data=(select blob2clobasc(data) from my_contents_lob b where a.pk1= b.pk1)
        where pk1 in (select al.pk1 from my_contents_lob al where dbms_lob.compare(blob2clob(al.data),a.data) =0 );
    SQL> select dump(dbms_lob.substr(data,4000,1)) from my_contents where pk1 in (select pk1 from my_contents_lob);
    Confirms that we're now working properly.
    To run across all the _LOB tables we've created:
    [oracle@RESTORECLONE ~]$ exp file=all_fixed_lobs.dmp buffer=4000000 userid=my_user/mypass tables=MY_CONTENTS_LOB,MY_FORUM_LOB...
    [oracle@RESTORECLONE ~]$ scp all_fixed_lobs.dmp jboulier@PRODCLONE:/tmp
    And then on PRODCLONE we can import:
    imp file=all_fixed_lobs.dmp buffer=4000000 userid=system/XXXXXXX fromuser=my_user touser=my_user
    Instead of running the above update statement for all the affected tables, we can use a simple stored procedure:
    create or replace procedure fix_us7_CLOBS
      (TABLE_NAME varchar2,
         FIX_COL varchar2 )
        authid current_user
        as
         orig_sql varchar2(1000);
         bak_sql  varchar2(1000);
        begin
        dbms_output.put_line('Creating '||TABLE_NAME||'_PRECONV to preserve the original data in the table');
        bak_sql:='create table '||TABLE_NAME||'_preconv as select pk1,'||FIX_COL||' from '||TABLE_NAME||' where pk1 in (select pk1 from '||TABLE_NAME||'_LOB) ';
        execute immediate bak_sql;
        orig_sql:='update '||TABLE_NAME||' tabnew set '||FIX_COL||'= (select blob2clobasc ('||FIX_COL||') from '||TABLE_NAME||'_LOB taborig where tabnew.pk1=taborig.pk1)
       where pk1 in (
       select a.pk1 from '||TABLE_NAME||'_LOB a,'||TABLE_NAME||' b
          where a.pk1=b.pk1
                 and dbms_lob.compare(blob2clob(a.'||FIX_COL||'),b.'||FIX_COL||') = 0 )';
        -- dbms_output.put_line(orig_sql);
        execute immediate orig_sql;
       end;
    Now we can run the procedure and it fixes everything for our previously-broken tables, keeping the changed rows -- just in case -- in a table called table_name_PRECONV.
    set serveroutput on time on timing on;
    exec fix_us7_clobs('MY_CONTENTS','DATA');
    commit;
    After confirming with the client that the changes work -- and haven't noticeably broken anything else -- the same routines can be carefully run against the actual production database.

    We converted using the database using scripts I developed. I'm not quite sure how we converted is relevant, other than saying that we did not use the Oracle conversion utility (not csscan, but the GUI Java tool).
    A summary:
    1) We replaced the lossy characters by parsing a csscan output file
    2) After re-scanning with csscan and coming up clean, our DBA converted the database to AL32UTF8 (changed the parameter file, changing the character set, switched the semantics to char, etc).
    3) Final step was changing existing tables to use char semantics by changing the table schema for VARCHAR2 columns
    Any specific steps I cannot easily answer, I worked with a DBA at our company to do this work. I handled the character replacement / DDL changes and the DBA ran csscan & performed the database config changes.
    Our actual error message:
    ORA-31011: XML parsing failed
    ORA-19202: Error occurred in XML processing
    LPX-00210: expected '<' instead of '�Error at line 1
    31011. 00000 - "XML parsing failed"
    *Cause:    XML parser returned an error while trying to parse the document.
    *Action:   Check if the document to be parsed is valid.
    Error at Line: 24 Column: 15
    This seems to match the the document ID referenced below. I will ask our DBA to pull it up and review it.
    Please advise if more information is needed from my end.

  • Special character

    Hi friends..
    Im getting the following error while ods activation.
    Value 'AMISH GANDHI#4604 ' (hex. '41004D004900530048002000470041004E0044004800490009') of characteristic 0POSTXT contains invalid.
    In ODS and PSA,i dont find # value,and instead of # there is a small square in the PSA and ODS and when i try to copy it in word i just get a space.
    What is that box like character?
    Im planning to use the following routine to control the special characters as given in one of the forum discussion with ALL_capital in RSKC.
    Will this solve my problem??
    Global Declaration
    DATA: L_S_ERRORLOG TYPE RSSM_S_ERRORLOG_INT,
    L_TEXT TYPE STRING,
    G_ALLOWED_CHAR(200) TYPE C.
    Routine
    RESULT = TRAN_STRUCTURE-MSEGNAME.
    Hash is not permitted at all
    while RESULT ca '#'.
    RESULT+sy-fdpos(1) = ' '.
    endwhile.
    Exclamation mark is not permitted as a first symbol of the field
    content
    if RESULT (1) = '!'.
    RESULT (1) = ' '.
    endif.
    *for characteristics to be used for navigation – lower case is not
    *permitted
    TRANSLATE RESULT TO upper case.
    *check if the list of permitted characters provided in the start routing
    *global variable
    *if not – derive it from configuration
    IF G_ALLOWED_CHAR IS INITIAL.
    CALL FUNCTION 'RSKC_ALLOWED_CHAR_GET'
    IMPORTING
    E_ALLOWED_CHAR = G_ALLOWED_CHAR.
    ENDIF.
    if allowed_char = ALL_CAPITAL, all capital letters are allowed,
    otherwise remove characters not specified in RSKC transaction
    CHECK G_ALLOWED_CHAR NE 'ALL_CAPITAL'.
    do.
    if not RESULT co g_allowed_char.
    shift RESULT+sy-fdpos left.
    else.
    exit.
    endif.
    enddo.
    returncode <> 0 means skip this record
    RETURNCODE = 0.
    abort <> 0 means skip whole data package!!!
    ABORT = 0.
    Thanks & Regards
    ragu

    Hi Ragu
    You need to use following transfer routine for this Infoobject in transfer rules. Then load the data in ODS and try activating data in ODS.
    You need to replace –XX with you transfer structure field.
    DATA: L_USER_ALLOWED_CHAR TYPE RSALLOWEDCHAR,
    L_ALL_ALLOWED_CHAR(140) TYPE C,
    L_RESULT_STR_LEN TYPE I,
    L_STR_INCREMENT TYPE I.
    CONSTANTS C_SAP_ALLOWED_CHAR(58) TYPE C VALUE
    ' !"%&''()*+,-./:;<=>?_0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZ'.
    IF L_ALL_ALLOWED_CHAR IS INITIAL.
    SELECT SINGLE * FROM RSALLOWEDCHAR
    INTO L_USER_ALLOWED_CHAR
    WHERE ALLOWKEY = 'S'.
    CONCATENATE C_SAP_ALLOWED_CHAR
    L_USER_ALLOWED_CHAR-ALLOWCHAR
    INTO L_ALL_ALLOWED_CHAR.
    ENDIF.
    RESULT = TRAN_STRUCTURE-XX.
    TRANSLATE RESULT TO UPPER CASE.
    L_RESULT_STR_LEN = STRLEN( RESULT ).
    L_STR_INCREMENT = 0.
    WHILE L_STR_INCREMENT <> L_RESULT_STR_LEN.
    IF NOT RESULT+L_STR_INCREMENT(1) CO L_ALL_ALLOWED_CHAR.
    RESULT+L_STR_INCREMENT(1) = ' '.
    ENDIF.
    ADD 1 TO L_STR_INCREMENT.
    ENDWHILE.
    Hope it helps.
    Regards
    Pradip

Maybe you are looking for

  • ORABPEL-10585 exception raised when starting the admin server

    Greetings, Have installed Oracle SOA Suite 11gR1 with WebLogic 10.3. Starting the admin application server raises the following exception: <Mar 30, 2012 5:30:33 PM CEST> <Error> <oracle.bpm.services.organization> <BEA-000000> <Exception exception.706

  • Showing Different amounts of Free space in DiskU than Macintosh HD

    On DiskU its showing that i have Used 86.76 GB, then the Macntosh HD is showning that i have only used 36.57 GB.Im only after re-installing Lion on my Macbook Pro. HOw can i fix this? Cheers in Advance

  • Microsoft Apps for Nokia N8

    Hi I got Nokia N8 (Product code 0599255, Software v: 111.030.0609 india). I have been tring for Microsoft apps for a long time but i still haven't got any update. Please halp.

  • Desktop Manager 4.3 fails to Launch

    Hi all Just upgraded to DTM 4.3, and it fails to show any icons on the screen just a blank white box. Currently running XP SP2 on my PC, had no problems with the older versions of DTM. I've tried removing it and installing it again, but the same prob

  • Pages 4.2 and Color Emoji Font

    Can it be true that the new Pages update still does not support the Apple Color Emoji font which Apple included with 10.7?  Does everyone else also find that the color versions will not display?