Loading and Writing Budget Data FDM vs ERPi

Say one wants to load data from a GL (EBS Suite) into Planning, and then write this budget data back to the EBS. Also need drill through functionality off planning web forms. What is the best way to do this?
I was thinking load data using FDM (so users can get the drill through) and then write back to EBS using ERPi. Does this sound right?

Thanks a lot for your response as always. Two questions more:
1)Can data be loaded using erpi? Or is it only for writing back?
2)If data can be loaded using erpi, then do I need fdm? If I load using erpi can i still get my drill through?

Similar Messages

  • Loading and Unloading of data

    Hi all,
    Today i have come accross the site http://www.wisdomforce.com/ in which the have a product named fast reader ,
    They claims that it can extract all the data's with in a fraction of second and never worried of it size...
    Anybody else using this product ..
    If yes,How this can be done in few seconds ,
    and Is there any similar product available from oracle apart from the traditional way of exp/imp.
    Regards,
    Mohammed

    It is not logical you can extract any size in fraction of second with Fastreader
    it will take more time to extract 1 billion rows tables compared to one million.
    Fastreader could improve the extract and load speed, especially if data has to be extracted into flat files
    http://www.associatedcontent.com/article/2081461/how_to_extract_and_load_data_from_oracle.html?cat=15

  • Question around UTL_FILE and writing unicode data to a file.

    Database version : 11.2.0.3.0
    NLS_CHARACTERSET : AL32UTF8
    OS : Red Hat Enterprise Linux Server release 6.3 (Santiago)
    I did not work with multiple language characters and manipulating them. So, the basic idea is to write UTF8 data as Unicode file using UTL_FILE. This is fairly an empty database and does not have any rows in at least the tables I am working on. First I inserted a row with English characters in the columns.
    I used utl_file.fopen_nchar to open and used utl_file.put_line_nchar to write it to the file on the Linux box. I open the file and I still see English characters (say "02CANMLKR001".
    Now, I updated the row with some columns having Chinese characters and ran the same script. It wrote the file. Now when I "vi" the file, I see "02CANè¹æ001" (some Unicode symbols in place of the Chinese characters and the regular English.
    When I FTP the file to windows and open it using notepad/notepad++ it still shows the Chinese characters. Using textpad, it shows ? in place of Chinese characters and the file properties say that the file is of type UNIX/UTF-8.
    My question : "Is my code working and writing the file in unicode? If not what are the required changes?" -- I know the question is little vague, but any questions/suggestions towards answering my question would really help.
    sample code:
    {pre}
    DECLARE
       l_file_handle   UTL_FILE.file_type;
       l_file_name     VARCHAR2 (50) := 'test.dat';
       l_rec           VARCHAR2 (250);
    BEGIN
       l_file_handle := UTL_FILE.fopen_nchar ('OUTPUT_DIR', l_file_name, 'W');
       SELECT col1 || col2 || col3 INTO l_rec FROM table_name;
       UTL_FILE.put_line_nchar (l_file_handle, l_rec);
       UTL_FILE.fclose (l_file_handle);
    END;
    {/pre}

    Regardless of what you think of my reply I'm trying to help you.
    I think you need to reread my reply because I can't find ANY relation at all between what I said and what you responded with.
    I wish things are the way you mentioned and followed text books.
    Nothing in my reply is related to 'text books' or some 'academic' approach to development. Strictly based on real-world experience of 25+ years.
    Unfortunately lot of real world projects kick off without complete information.
    No disagreement here - but totally irrevelant to anything I said.
    Till we get the complete information, it's better to work on the idea rather than wasting project hours. I don't think it can work that way. All we do is to lay a ground preparation, toy around multiple options for the actual coding even when we do not have the exact requirements.
    No disagreement here - but totally irrevelant to anything I said.
    And I think it's a good practice rather than waiting for complete information and pushing others.
    You can't just wait. But you also can't just go ahead on your own. You have to IMMEDIATELY 'push others' as soon as you discover any issues affecting your team's (or your) ability to meet the requirements. As I said above:
    Your problems are likely:
    1. lack of adequate requirements as to what the vendor really requires in terms of format and content
    2. lack of appropriate sample data - either you don't have the skill set to create it yourself or you haven't gotten any from someone else.
    3. lack of knowledge of the character sets involved to be able to create/conduct the proper tests
    If you discover something missing with the requirements (what character sets need to be used, what file format to use, whether BOMs are required in the file, etc) you simply MUST bring that to your manager's attention as soon as you suspect it might be an issue.
    It is your manager's job, not yours, to make sure you have the tools needed to do the job. One of those tools is the proper requirements. If there is ANYTHING wrong, or if you even THINK something is wrong with those requirements it is YOUR responsibility to notify your manager ASAP.
    Send them an email, leave a yellow-sticky on their desk but notify them. Nothing in what I just said says, or implies, that you should then just sit back and WAIT until that issue is resolved.
    If you know you will need sample data you MUST make sure the project plan includes SOME means to obtain sample data witihin the timeline needed by your project. As I repeated above if you don't have the skill set to create it yourself someone else will need to do it.
    I did not work with multiple language characters and manipulating them.
    Does your manager know that? If the project requires 'work with multiple language characters and manipulating them' someone on the project needs to have experience doing that. If your manager knows you don't have that experience but wants you to proceed anyway and/or won't provide any other resource that does have that experience that is ok. But that is the manager's responsibility and that needs to be documented. At a minimum you need to advise your manager (I prefer to do it with an email) along the following lines:
    Hey - manager person - As you know I have little or no experience to 'work with multiple language characters and manipulating them' and those skills are needed to properly implement and test that the requirements are met. Please let me know if such a resource can be made available.
    And I'm serious about that. Sometimes you have to make you manager do their job. That means you ALWAYS need to keep them advised of ANY issue that might affect the project. Once your manager is made aware of an issue it is then THEIR responsibility to deal with it. They may choose to ignore it, pretend they never heard about it or actually deal with it. But you will always be able to show that you notified them about it.
    Now, I updated the row with some columns having Chinese characters and ran the same script.
    Great - as long as you actually know Chinese that is; and how to work with Chinese characters in the context of a database character set, querying, creating files, etc.
    If you don't know Chinese or haven't actually worked with Chinese characters in that context then the project still needs a resource that does know it.
    You can't just try to bluff your way through something like character sets and code conversions. You either know what a BOM (byte order mark) is or you don't. You have either learned when BOMs are needed or you haven't.
    That said, we are in process of getting the information and sample data that we require.
    Good!
    Now make sure you have notified your manager of any 'holes' in the requirements and keep them up to date with any other issues that arise.
    NONE of the above suggests, or implies, that you should just sit back and wait until that is done. But any advice offered on the forums about specifics of your issue (such as whether you need to even worry about BOMs) is premature until the vendor or the requirements actually document the precise character set and file format needed.

  • (Urgent) form taking lots of time to load and fetch the data

    Hi
    I have very serious problem of perfomance. I have installed oracle portal and configured it properly.
    Now except form everything working perfectly. but only forms and links are taking lots of time to load
    as well as for fetching data. I try to tune my sga and aslo findout hits and miss but i didnt find any
    any problem for it. it is ok. can u help me pls. how can i make my form fast.
    My operation system is NT.
    version of database oracle 8.1.7
    Oracle Portal 3.0.8
    Thanks in advance
    raju parmar
    [email protected]
    null

    hi chetan
    Lots of Thanks.
    Now it working perfect.
    Raju
    <BLOCKQUOTE><font size="1" face="Verdana, Arial">quote:</font><HR>Originally posted by Chetan Kashyap ([email protected]):
    The workaround is provided at: http://technet.oracle.com:89/ubb/Forum81/HTML/000395.html <HR></BLOCKQUOTE>
    null

  • Problems with parport and writing to data adresse 0x037a

    Hi,
    Iam currently trying to control a DC-motor, I need to change the strobe which is at the data control. At adresss 0x037a i cant write to it. Running userport on winxp and i can access and turn data ports at 0x0378 on and off.
    I get this error message:
    XCEPTION_PRIV_INSTRUCTION (0xc0000096) at pc=0x1000107b, pid=3584, tid=1064
    Code:
    import parport.ParallelPort;
    class SimpleIO {
    public static void main ( String []args )
    //ParallelPort lpt1 = new ParallelPort(0x378); // 0x378 is normally the base address for the LPT1 port
    //int aByte;
    //aByte = lpt1.read(); // read a byte from the port's STATUS pins
    //System.out.println("Input from parallel port: " + aByte);
    //int a=0xFF;
    //lpt1.write(a); // write a byte to the port's DATA pins
    ParallelPort lpt2 = new ParallelPort(0x03bc);
    //lpt2.write(0x00);
    int c=0x0379;
    //System.out.println("Output to port: " + aByte);
    //ParallelPort lpt2 = new ParallelPort(0x37A);
    //int b=0x00;
    //lpt2.write(b);
    }

    Hola!!!
    Me preguntaba de donde sacaste esa APi de parport.ParallelPort ???
    Precisamente yo tambien quiero usar el puero paralelo para controlar un motor a pasos, pero no lo logro conseguir, me baje la API de javax.comm y todo parece estar bien solo que no obtengo nada a la salida del puerto. Creo que es porque WinXP te cierra los puertos, no sabes como abrirlos???
    Mi codigo es el siguiente:
    import java.io.*;
    import javax.comm.*;
    public class TestPuertoParalelo{
    public static void main(String[] args){
    CommPortIdentifier portId;
    ParallelPort paralelo = null;
    InputStream entrada = null;
    OutputStream salida = null;
    try{
    portId = CommPortIdentifier.getPortIdentifier("LPT1");
    paralelo = (ParallelPort) portId.open("ParallelPort",5000);
    paralelo.setMode(ParallelPort.LPT_MODE_SPP); //parece bien SPP, los demas modos no me funcionan
    salida = paralelo.getOutputStream();
    System.out.println ("escribiendo 8");
    salida.write(8); //Aqui se queda y no recibo nada en el puerto
    System.out.println("Se Recibio Correctamente");
    catch(NoSuchPortException e){
    System.out.println("Error del Id Puerto");
    catch(PortInUseException e) {
    System.out.println("Error: Puerto en Uso");
    catch(UnsupportedCommOperationException e){
    System.out.println("Error en el Modo");
    catch(IOException e){
    System.out.println("Error en la conexion de flujos");
    }

  • Dbwr and writing undo data block

    hi guys,
    If dbwr was to write a dirty block to disk, would it also have to write the dirty uno data block that contains the before value of our database block to disk too?
    Thanks

    OracleGuy777 wrote:
    hi guys,
    If dbwr was to write a dirty block to disk, would it also have to write the dirty uno data block that contains the before value of our database block to disk too?
    The first thing to remember is that recovery depends on the REDO, not the UNDO, so technilcally it doesn't matter whether dbwr manages to write neither, one, or both of the two blocks to disc to avoid any problems with crash recovery. This should give you a clue that dbwr doesn't HAVE to write the undo block at the same time as the data block.
    In practice, it is possible that both blocks will be written virtually simultaneously because the two blocks will have been modified at the same instant and therefore could be written during the same 3-second timeout write (continuous checkpoint). It is also possible, though, for the undo block to be written some time before the data block, or some time after the data block - it all depends on what activity has been going at about the same time.
    Regards
    Jonathan Lewis
    http://jonathanlewis.wordpress.com
    http://www.jlcomp.demon.co.uk
    To post code, statspack/AWR report, execution plans or trace files, start and end the section with the tag {noformat}{noformat} (lowercase, curly brackets, no spaces) so that the text appears in fixed format.
    "Science is more than a body of knowledge; it is a way of thinking"
    Carl Sagan                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

  • Record count is different for Full load and Init with data transfer

    Hi all,
    We have a data source for which delta is enabled on Calendar Day.
    If I run the full load record count is 2670.But If I run the delta record count is 2665.5 records are missing.
    Would you please help on this.Why those records are missing.

    Hi,
    Actually for Full Load say you have 50 records. When you run it will upload 50 records.
    Now if you run delta if there is no records changes it wouldnt update because there is no update in the existing records.
    only delta is done when there is changes in the records.

  • Reading the Blob and writing it to an external file in an xml tree format

    Hi,
    We have a table by name clarity_response_log and content of the column(Response_file) is BLOB and we have xml file or xml content in that column. Most probably the column or table may be having more than 5 records and hence we need to read the corresponding blob content and write to an external file.
    CREATE TABLE CLARITY_RESPONSE_LOG
      REQUEST_CODE   NUMBER,
      RESPONSE_FILE  BLOB,
      DATE_CRATED    DATE                           NOT NULL,
      CREATED_BY     NUMBER                         NOT NULL,
      UPDATED_BY     NUMBER                         DEFAULT 1,
      DATE_UPDATED   VARCHAR2(20 BYTE)              DEFAULT SYSDATE
    )The xml content in the insert statement is very small because of some reason and cannot be made public and indeed we have a very big xml file stored in the BLOB column or Response_File column
    Insert into CLARITY_RESPONSE_LOG
       (REQUEST_CODE, RESPONSE_FILE, DATE_CRATED, CREATED_BY, UPDATED_BY, DATE_UPDATED)
    Values
       (5, '<?xml version="1.0" encoding="UTF-8"?><xml-response><phone-number>1212121212</tracking-number></xml-response>', TO_DATE('09/23/2010 09:01:34', 'MM/DD/YYYY HH24:MI:SS'), 1, 1, '23-SEP-10');
    Insert into CLARITY_RESPONSE_LOG
       (REQUEST_CODE, RESPONSE_FILE, DATE_CRATED, CREATED_BY, UPDATED_BY, DATE_UPDATED)
    Values
       (6, '<?xml version="1.0" encoding="UTF-8"?><xml-response><phone-number>1212121212</tracking-number></xml-response>', TO_DATE('09/23/2010 09:01:34', 'MM/DD/YYYY HH24:MI:SS'), 1, 1, '23-SEP-10');
    Insert into CLARITY_RESPONSE_LOG
       (REQUEST_CODE, RESPONSE_FILE, DATE_CRATED, CREATED_BY, UPDATED_BY, DATE_UPDATED)
    Values
       (7, '<?xml version="1.0" encoding="UTF-8"?><xml-response><phone-number>1212121212</tracking-number></xml-response>', TO_DATE('09/23/2010 09:01:34', 'MM/DD/YYYY HH24:MI:SS'), 1, 1, '23-SEP-10');
    Insert into CLARITY_RESPONSE_LOG
       (REQUEST_CODE, RESPONSE_FILE, DATE_CRATED, CREATED_BY, UPDATED_BY, DATE_UPDATED)
    Values
       (8, '<?xml version="1.0" encoding="UTF-8"?><xml-response><phone-number>1212121212</tracking-number></xml-response>', TO_DATE('09/23/2010 09:01:34', 'MM/DD/YYYY HH24:MI:SS'), 1, 1, '23-SEP-10');
    Insert into CLARITY_RESPONSE_LOG
       (REQUEST_CODE, RESPONSE_FILE, DATE_CRATED, CREATED_BY, UPDATED_BY, DATE_UPDATED)
    Values
       (9, '<?xml version="1.0" encoding="UTF-8"?><xml-response><phone-number>1212121212</tracking-number></xml-response>', TO_DATE('09/23/2010 09:01:34', 'MM/DD/YYYY HH24:MI:SS'), 1, 1, '23-SEP-10');THe corresponding proc for reading the data and writing the data to an external file goes something like this
    SET serveroutput ON
    DECLARE
       vstart     NUMBER             := 1;
       bytelen    NUMBER             := 32000;
       len        NUMBER;
       my_vr      RAW (32000);
       x          NUMBER;
       l_output   UTL_FILE.FILE_TYPE;
    BEGIN
    -- define output directory
       l_output :=
          UTL_FILE.FOPEN ('CWFSTORE_RESPONCE_XML', 'extract500.txt', 'wb', 32760);
       vstart := 1;
       bytelen := 32000;
    ---get the Blob locator
       FOR rec IN (SELECT response_file vblob
                     FROM clarity_response_log
                    WHERE TRUNC (date_crated) = TRUNC (SYSDATE - 1))
       LOOP
    --get length of the blob
    len := DBMS_LOB.getlength (rec.vblob);
          DBMS_OUTPUT.PUT_LINE (len);
          x := len;
    ---- If small enough for a single write
    IF len < 32760
          THEN
             UTL_FILE.put_raw (l_output, rec.vblob);
             UTL_FILE.FFLUSH (l_output);
          ELSE  
    -------- write in pieces
             vstart := 1;
             WHILE vstart < len AND bytelen > 0
             LOOP
                DBMS_LOB.READ (rec.vblob, bytelen, vstart, my_vr);
                UTL_FILE.put_raw (l_output, my_vr);
                UTL_FILE.FFLUSH (l_output);
    ---------------- set the start position for the next cut
                vstart := vstart + bytelen;
    ---------- set the end position if less than 32000 bytes
                x := x - bytelen;
                IF x < 32000
                THEN
                   bytelen := x;
                END IF;
                UTL_FILE.NEW_LINE (l_output);
             END LOOP;
    ----------------- --- UTL_FILE.NEW_LINE(l_output);
          END IF;
       END LOOP;
       UTL_FILE.FCLOSE (l_output);
    END;The above code works well and all the records or xml contents are being written simultaneously adjacent to each other but we each records must be written to a new line or there must be a line gap or a blank line between any two records
    the code which I get is as follow all all xml data comes on a single line
    <?xml version="1.0" encoding="ISO-8859-1"?><emp><empno>7369</empno><ename>James</ename><job>Manager</job><salary>1000</salary></emp><?xml version="1.0" encoding="ISO-8859-1"?><emp><empno>7370</empno><ename>charles</ename><job>President</job><salary>500</salary></emp>But the code written to an external file has to be something like this.
    <?xml version="1.0" encoding="ISO-8859-1"?>
    <emp>
      <empno>7369</empno>
      <ename>James</ename>
      <job>Manager</job>
      <salary>1000</salary>
    </emp>
    <?xml version="1.0" encoding="ISO-8859-1"?>
    <emp>
    <empno>7370</empno>
    <ename>charles</ename>
    <job>President</job>
    <salary>500</salary>
    </emp>Please advice

    What was wrong with the previous answers given on your other thread:
    Export Blob data to text file(-29285-ORA-29285: file write error)
    If there's a continuing issue, stay with the same thread, don't just ask the same question again and again, it's really Pi**es people off and causes confusion as not everyone will be familiar with what answers you've already had. You're just wasting people's time by doing that.
    As already mentioned before, convert your BLOB to a CLOB and then to XMLTYPE where it can be treated as XML and written out to file in a variety of ways including the way I showed you on the other thread.
    You really seem to be struggling to get the worst possible way to work.

  • Reading/Writing Bulk Data, getting OutOfMemoryError: Java heap space

    Hello,
    I am reading more than a million record using CursoredStream, following is my code snippet for reading data
    ReadAllQuery query = new ReadAllQuery();
    query.setReferenceClass(DataBk.class);
    query.useCursoredStream();
    CursoredStream stream;
    stream = (CursoredStream)dbses.executeQuery(query);
           while (!stream.atEnd()) {
           vector1 = stream.read(stream.size());
           }       Problem is I am getting
    Exception in thread "main" java.lang.OutOfMemoryError: Java heap spaceI was reading other forums posts about this and it is mentioned that memory heap size could be increased.
    Where do I increase memory heap size?
    java.lang.OutOfMemoryError while opening 9.0.4.5 mwp in 10.1.3.3
    Besides, if my above code is not alright, someone could suggest best approach for reading and writing bulk data.
    Regards
    Edited by: user20090209 on Oct 14, 2009 10:17 AM

    Hello,
    If you are reading a million records and running out of heap, you will still get the exception by reading it using a cursor one at a time. It also depends on how the driver is processing the million records. Using a scrollable cursor might help ( http://wiki.eclipse.org/Using_Advanced_Query_API_(ELUG) ), increasing the heap as suggested will (In Jdev, Java options can be added to the project properties->Run/Debug/Profile->Launch settings), or even changing your app design to use objects more efficiently all around. It depends on what your are using the objects for, but even using pagination (via setting the firstRows/maxResults on a query) to limit the results returned can help depending on how the application requires the objects. What is important is that the objects be free for garbage collection when you are done processing them, and before the next batch needs to be brought in from the database.
    Also, common problems are that reading in one object can bring in multiple trees if relationships aren't managed efficiently. Only use relationships where required, and use indirection where ever possible.
    Best Regards,
    chris

  • Extraction and loading of Training and Event Management data (0HR_PE_1)

    hello,
    I've got the following doubt:
    before BI 7.0 release, extractor 0HR_PE_1 extracts event data (eventid, attendee id, calday,...) but if you load straight to cube 0PE_C01, as Calendar year/month needs a reference date (for example, event start date), you'll get total duration of event in hours or days refered to event star date.
    So in a query filtered by month, you get total duration of an event that starts in the filtered month but it couldn`t end until few months later o year later, so you don´t get appropiate information.
    Example:
    Event          calday        Hours
    10004377  20081120   500        but from event_attr event end date is 20090410.
    In a query filtered by 200811 you get total duration time (500 hours when in 200811 event hours have been 20) or if you filter by any month of 2009 you don´t get information of duration of that event.
    I had to create a copy of standar cube (without calday, only Calendar year/month, Calendar year in time dimension) and disaggrate data creating as many entries for an event as months lasts and adjust calculation of ratios duration of event (duration of event in each month/ total duration of event).
    The question is: Is there any improvement or change on business content in BI 7.0 to treat Training and Event Management data? Has anybody had to deal with that?
    Thank you very much.
    IRB

    Hi,
    TEM data is stored in HRP tables.
    You can load the catalog by creating LSMWs for objects Business event group (L), Business event types (D), Locations (F), Organizers (U) as per requirement.
    LSMW for tcode PP01 can be used to create these objects.
    To create Business Events (E) you can create LSMW for PV10/PV11.
    To book attendee create LSMW for tcode PV08 as here you can specify the actual business event ID which reduces ambiguity.
    tcode PV12 to firmly book events
    tcode PV15 to follow up
    Hope this helps.
    Regards,
    Shreyasi.

  • Training and Event Management Data Load

    Hello Team
    Would appreciate if any of you can advice on how to load Training and Event management data. I think its stored in HRP tables.
    I'm working on an upgrade assignment.
    Thanks

    Hi,
    TEM data is stored in HRP tables.
    You can load the catalog by creating LSMWs for objects Business event group (L), Business event types (D), Locations (F), Organizers (U) as per requirement.
    LSMW for tcode PP01 can be used to create these objects.
    To create Business Events (E) you can create LSMW for PV10/PV11.
    To book attendee create LSMW for tcode PV08 as here you can specify the actual business event ID which reduces ambiguity.
    tcode PV12 to firmly book events
    tcode PV15 to follow up
    Hope this helps.
    Regards,
    Shreyasi.

  • My iPhone 4S will not connect to the internet properly. It says that it is connected, the App Store won't load apps and I can't get on my face book app. I have tried wifi and my cellular data, neither work. What are some ways to fix this?

    My iPhone will
    not connect to the internet properly. It says that it is connected, the App Store won't load apps and I can't get on my face book app. I have tried wifi and my cellular data, neither work. What are some ways to fix this?

    Try resetting your network settings.
    Settings>Genenal>Reset>Reset Network Settings.

  • Safari won't open, just quits before loading. I'm on Yosimite and up to date.

    Safari won't open, just quits before loading. I'm on Yosimite and up to date.

    Do a backup.
    Go to Finder and select your user/home folder. With that Finder window as the front window, either select Finder/View/Show View options or go command - J.  When the View options opens, check ’Show Library Folder’. That should make your user library folder visible in your user/home folder.  Select Library. Then go to Preferences/com.apple.Safari.plist. Move the .plist to your desktop.
    Restart, open the application and test. If it works okay, delete the plist from the desktop.
    If the application is the same, return the .plist to where you got it from, overwriting the newer one.
    Thanks to leonie for some information contained in this.

  • ICal reports that it can't find server, and lots of data from several years ago is trying to get attention.  Cannot bring up calendar in any view to use. also keeps loading unused calendar websites from long ago.  I can' remove them. what to do?

    iCal reports that it can't find server, and lots of data from several years ago is trying to get attention.  Cannot bring up calendar in any view to use. also keeps loading unused calendar websites from long ago.  I can' remove them. what to do?

    UPDATE: I had 36 Safari windows open and minimized on the dock from my last session; 20 of those windows failed to open b/c Safari "couldn't find the server"; I just went to the URL bar of all 20 pages and clicked return, and all 20 pages loaded perfectly, without dropping any images or losing formatting. This is a puzzle!!!

  • Hi, I set up my new iPhone S4 with iclouds. So the data of my old S3 should move to the S4. However, some applications could not be loaded on the S04. Now some apps are loaded and others not . How can I progress?

    Hi, I set up my new iPhone S4 with iclouds. So the data of my old S3 should move to the S4. However, some applications could not be loaded on the S04. Now some apps are loaded and others not . How can I progress?

    Thanks. I can change the forwarding settings in Gmail to be either POP or IMAP. That is quite simple. I just looked at those settings on the Gmail server and, to my surprise, both are enabled. In order for outlook client on my computer to receive IMAP emails I have to change the email account settings in my current outlook from POP to IMAP. I don't need to be using the Exchange version of Outlook to receive IMAP. I guess I'll do a little research tomorrow on Exchange so i can figure out what, if anything, it might do for me.
    I glanced at the link you sent me and it looks like it might be very helpful - another project for tomorrow.
    BTW: I hit "This Solved my Question" by accident. I meant only to "like' your response so you would get some benefit from the help you've given me so far. I hope I haven't discouraged others from taking a shot at helping me. I suspect there is more than one way to approach this and, as you know, my problem isn't really solved until I have what I need all set up and running smoothly.

Maybe you are looking for

  • File sharing across a network --- Mac (Leopard) and PCs (Vista) -- problems

    Been trying to setup file sharing so that I can access an iMac from a Vista PC. Followed various help threads, and have a specific question as relates to setting the file sharing preferences. From system preferences/sharing, have checked "File Sharin

  • HTTP Sender Adapter - INTERFACE_REGISTRATION_ERROR

    Hi All, I am getting this error when I am trying to call a synchronous interface using Plain HTTP service. The interface then calls RFC in SAP system which is returned to The interface is defined and activated in Integration Repository and Receiver a

  • Anyone have no service pop up with the new update and stay like that?

    Anyone have no service pop up with the new update and stay like that

  • Blurry text in Adobe Content Viewer

    I have a problem with my resolution, I think. When I look at my folio in the Content Viwer on my iPad the text is blurry and out of focus. Why is that? Am I doing anything wrong or is it because I haven´t published it yet? Is the preview lowres? Plea

  • Rule based ATP check

    Hello, We are implementening rule based ATP check for Sales order scnerio. Our Business process is 1.Customer will cretate order for product P1 & Location L1. If stock is available, system will confirm the order else it should search for alternate pr