Values in Colum as separate columns

I have a table that looks like this:
Company (varchar),
FieldName (varchar)
Valuex (varchar)
Data may look like this:
Company Fieldname Valuex
123 Name John
123 MiddleName C
123 LastName Smith
I would like to write a query that comes out like this:
Company FirstName MiddleName LastName
123 John C Smith
Can it be one? Thanks in advance.

Hi,
Pivot is new in 11g is it not? I'm on 10g. Can it be done without the pivot function?Yes! You can do pivots in any version. I just said there was a newer, easier way that was introduced in Oracle 11; that's the only case in which the word PIVOT occurs in the code.
I just tried searching for "pivot", and I did find dozens of examples. Almost all of them said "Search for pivot", so I'll post an exmple here. The example below works in any version of Oracle, including 10.
--     How to Pivot a Result Set (Display Rows as Columns)
--     This example uses the scott.emp table.
--     Given a query that produces three rows for every department,
--     how can we show the same data in a query that has one row
--     per department, and three separate columns?
--     For example, the query below counts the number of employess
--     in each departent that have one of three given jobs:
PROMPT     ==========  0. Simple COUNT ... GROUP BY  ==========
SELECT     deptno
,     job
,     COUNT (*)     AS cnt
FROM     scott.emp
WHERE     job     IN ('ANALYST', 'CLERK', 'MANAGER')
GROUP BY     deptno
,          job;
Output:
    DEPTNO JOB              CNT
        20 CLERK              2
        20 MANAGER            1
        30 CLERK              1
        30 MANAGER            1
        10 CLERK              1
        10 MANAGER            1
        20 ANALYST            2
PROMPT     ==========  1. Pivot  ==========
SELECT     deptno
,     COUNT (CASE WHEN job = 'ANALYST' THEN 1 END)     AS analyst_cnt
,     COUNT (CASE WHEN job = 'CLERK'   THEN 1 END)     AS clerk_cnt
,     COUNT (CASE WHEN job = 'MANAGER' THEN 1 END)     AS manager_cnt
FROM     scott.emp
WHERE     job     IN ('ANALYST', 'CLERK', 'MANAGER')
GROUP BY     deptno;
--     Output:
    DEPTNO ANALYST_CNT  CLERK_CNT MANAGER_CNT
        30           0          1           1
        20           2          2           1
        10           0          1           1
--     Explanation
(1) Decide what you want the output to look like.
     (E.g. "I want a row for each department,
     and columns for deptno, analyst_cnt, clerk_cnt and manager_cnt)
(2) Get a result set where every row identifies which row
     and which column of the output will be affected.
     In the example above, deptno identifies the row, and
     job identifies the column.
     Both deptno and job happened to be in the original table.
     That is not always the case; sometimes you have to
     compute new columns based on the original data.
(3) Use aggregate functions and CASE (or DECODE) to produce
     the pivoted columns. 
     The CASE statement will pick
     only the rows of raw data that belong in the column.
     If each cell in the output corresponds to (at most)
     one row of input, then you can use MIN or MAX as the
     aggregate function.
     If many rows of input can be reflected in a single cell
     of output, then use SUM, COUNT, AVG, STRAGG, or some other
     aggregate function.
     GROUP BY the column that identifies rows.
{code}
Edited by: Frank Kulash on Sep 5, 2008 5:37 PM                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

Similar Messages

  • Pivoting string values( ,comma seperator) to separate columns(usingREG_EXP)

    <pre>
    I have column PART which contains string as shown below
    PART
    BMW PART,TEST PART,TEST PART,HYPER PART,HYPER PART,TESTINGNEWONE,TESTINGNEWONE,TESTINGNEWONE
    string has got comma separated, i need to project as individual columns like
    P1 P2 .... P8
    BMW PART TEST PART TESTINGEWONE
    how can it be done... (i think by using regular expressions it can be done but i am not aware of it can any one help)
    </pre>
    Thanks
    Naren

    Hi,
    Narendar.Vishwanatham wrote:
    thanks a lot for your query....
    cant we make it dynamically instead of hardcoding each columnThe number of columns produced by a query must be hard-coded into the query.
    Dynamic SQL means that you have some program (often another query) write all or part of the main query for you.
    The following thread has an example of a dynamic pivot in SQL*Plus:
    Help for a query to add columns
    The following thread discusses different options for dealing with an unknown number of columns after the pivot:
    Re: Report count and sum from many rows into many columns
    (and i guess the answer you gave is not possible to get with out using regular expresssions right !!!)No, you can get the same results using INSTR and SUBSTR.
    Whenever you ask for help, post a little sample data (CREATE TABLE and INSERT statements), and the results you want from that data.
    Always mention what version of Oracle (and any other relevant tools) you're using. This is especially important with pivots; recent versions of Oracle have helpful new features.

  • How to extract member, default alias and custom alias into separate columns

    Hi,
    is there a way to get simultaneously member name, default alias name and custom alias name with Essbase 9x report scripts (Report Writer)?
    I do know that using for example
    <OUTALTSELECT "default or custom alias table"
    <REPMBRALIAS "dimension"
    I can get both the real member name and selected alias table, but the problem is that they're both returned to same column.
    Using fix columns and order I can repeat member name into several columns if needed, but I would need to tell what to put for each column.
    For example : { ORDER 0,0,0,1,2,3,4,5 FIXCOLUMNS 8 }
    I would like to force the report script to put Member name into first column in row, default alias name of member into second column in row and custom alias name of member into third column in row, like :
    MEMBERNAME DEFAULTALIASNAME CUSTOMALIASNAME
    Any ideas how to handle this?
    I can of course create several extracts containing one sort of info per list, but it would make life somewhat easier having all of these in same extract.
    Thanks in advance!
    Br,
    MK

    Hi,
    well, that's pretty much what I want to achieve, to get all info of member within same row, but into separate columns.
    The other option is to then produce 2 other extracts containing:
    Extract 1:
    MEMBERNAME DEFAULTALIASNAME
    and
    Extract 2:
    MEMBERNAME CUSTOMALIASNAME
    These values should be in separate columns?
    I think I lost the track while playing around with report writer, does anyone have a ready made solution for this?
    <REPMBRALIAS brings the needed info, but returns the values into same column.
    Empty space delimiter could be handled with customalias table, but in this case default alias name may have long text with several empty spaces making it hard to sort in the other system side.
    Any suggestions?
    Thanks so far for your kind help.
    Br,
    MK

  • Getting the values in separate columns

    Hello Everyone
    I am new to Bex, i have the sales data of 2008,2009,2010. now my requirement is to display the sales order(key figure) for 2008,2009,2010 in separate columns for each customer(dimension) in a single report , can any one help me how to get this done in bex.
    Thanks
    Gupta

    Hi Rohit,
    To do that you should have calyear in dimennsion, if so simple you can do it else do the following.
    Edit your cube add calyear in time dimension and map it accordingly. 
    Go To Bex Designer :
    Select required free char, row and column (KF),
    Drag and drop calyear on top of kf (Extreem top to KF). 
    when you run your query it will display all the value according to relavant year. 
    Note :  Suppose you have 4kf in your colum and having three years data in the target then there will be 12 columns will be displayed in report, ie number of kf  X  number of year
    So, for better view or perfomance please do not forget to give restriction of year by giving variable to user. 
    Thanks
    BVR

  • How to get the values in separate columns

    Hello Everyone
    I am new to Bex, i have the sales data of 2008,2009,2010. now i have to display the sales order(key figure) for 2008,2009,2010 in separate columns for each customer(dimension) in a single report , can any one help me how to get this done in bex.
    Thanks
    Gupta

    Hi ,
    You can achieve this either creating New Selections or Restricted Key Figures.
    Right click on the structute>>>> New selcection>>> drag the year to right [  Right Click on it >>restrict it with 2008}>>>Drag the respected Key figure.
    So u will get the values 2008 in seperate column.
    COpy the New Selecten that u created and paste on the structure , chage the descriprion and year to 2009.
    do the same for 2010.
    Note:you can also achieve this by creating variable offsets.Check the following link
    http://help.sap.com/saphelp_bw33/helpdata/en/3f/89533e5ff4d064e10000000a114084/content.htm
    Regards,
    Ranganath

  • I want to spit cell the value into two separate columns like f_name & l_nam

    Hi Guys,
    I have excel data in one cell with ',' separated data.
    sample data:
    empno ename
    121 ravi,kann
    232 ram,raman
    here ename value in one provided in one cell.
    Now i want to spit cell the value into two separate columns like first_name and last_name?
    Thanks in advance
    -LK

    with sample_data as
      select 121 empno, 'ravi,kann' ename from dual union all
      select 232 empno, 'ram,raman' from dual
    select empno,
           substr(ename,1,instr(ename,',')-1) first_name,
           substr(ename,instr(ename,',')+1) last_name
    from sample_data;       or with regular expressions:
    with sample_data as
      select 121 empno, 'ravi,kann' ename from dual union all
      select 232 empno, 'ram,raman' from dual
    select empno,
           regexp_substr(ename,'^[^,]*') first_name,
           regexp_substr(ename,'[^,]*$') last_name
    from sample_data; Edited by: hm on 04.07.2012 06:22

  • BO: separate column for + - sign apart from column with value

    Hello
    In Business Objects Web Intelligence I would like to create a column that shows the sign '+' or '-' when a key figure KF1 is below 0 or above. How do I do this? It has to be a separate column. I already created a calculated key figure KF2 that has as formula ABS (KF1). KF1 itself must not be shown as contains amount and sign in one field. I need to separate these 2.
    Regards
    Arjan

    You can create one more variable with the formula:
    =If([Key FIG 1]<0) Then "-" Else "+"
    Regards,
    Rohit

  • Convert mixed case value in colum into upper case values

    Hi All,
    I have got a table call emp and has the following colums:
    id number 10
    suename_name varchar2(20)
    given_name varchar2(20)
    position varchar2(30)
    date_hired date
    Now the values in colums are mixed case and I need to change them to upper case. I think I need to use upper function to do it.
    Could anyone help me by providing me sql/pl sql script to through each colum in emp table and change them to the upper case.
    Many Thanks
    Michael

    Hi,
    If you really want to try this using PL/SQL, then you'll have to use dynamic SQL, something like this untested code, since the table and column names are vaiable:
    SET   SERVEROUTPUT  ON  SIZE 50000
    DECLARE
         sql_txt     VARCHAR2 (4000);
    BEGIN
         FOR  r  IN ( SELECT  table_name
                   ,       column_name
                   FROM    user_tab_columns
                   WHERE   data_type = 'VARCHAR2'
              --   AND     ...  -- if you don't want all columns in all tables
         LOOP
              sql_txt := 'UPDATE "'
                   || r.table_name
                   || '" SET "'
                   || r.column_name
                   || '" = UPPER ("'
                   || r.column_name
                   || '")';
              dbms_output.put_line (sql_txt || ' = sql_txt');     -- For debugging
              -- EXECUTE IMMEDIATE sql_txt;               -- For real
         END LOOP;
    END;Dynamic SQL (here) means creating a string that contains the SQL statement to be done, then using EXECUTE IMMEDIATE to run it.
    Whenever you write dynamic SQL, it's a good idea to only display the SQL statement during the early stages of debugging. When it's displaying correctly, then you can try un-commenting the EXECUTE IMMEDIATE line.
    Edited by: Frank Kulash on Jul 19, 2009 4:40 PM
    A little more complicated, but a lot more efficient:
    SET   SERVEROUTPUT  ON  SIZE 50000
    DECLARE
         sql_txt          VARCHAR2 (4000);
    BEGIN
         FOR  r  IN ( SELECT    table_name
                   ,         column_name
                   ,         ROW_NUMBER () OVER ( PARTITION BY  table_name
                                      ORDER BY       column_name
                                    )     AS column_num
                   ,            COUNT (1)     OVER ( PARTITION BY  table_name
                                    )     AS total_column_cnt
                   FROM      user_tab_columns
                   WHERE     data_type = 'VARCHAR2'
              --   AND       ...  -- if you don't want all columns in all tables
                   ORDER BY  table_name
                   ,            column_name
         LOOP
              IF  r.column_num = 1
              THEN
                   sql_txt := 'UPDATE "'
                   || r.table_name
                   || '" SET "';
              ELSE
                   sql_txt := sql_txt || ', "';
              END IF;
              sql_txt := sql_txt ||
                   || r.column_name
                   || '" = UPPER ("'
                   || r.column_name
                   || '")';
              IF  r.column_num = r.total_column_cnt
              THEN     -- This is the last row for this table; run it
                   dbms_output.put_line (sql_txt || ' = sql_txt');     -- For debugging
                   -- EXECUTE IMMEDIATE sql_txt;               -- For real
              END IF;
         END LOOP;
    END;
    {code}
    The difference is that the first solution produces and executes a separate UPDATE statement for each column, like this:
    {code}
    UPDATE "TABLE_1" SET "COLUMN_A" = UPPER ("COLUMN_A");
    UPDATE "TABLE_1" SET "COLUMN_B" = UPPER ("COLUMN_B");
    UPDATE "TABLE_1" SET "COLUMN_C" = UPPER ("COLUMN_C");
    {code}
    but it's much more efficient to do change all the columns at once, as long as you have the row in hand.  So the second solution only creates one SQL statement per table, like this:
    {code}
    UPDATE "TABLE_1" SET "COLUMN_A" = UPPER ("COLUMN_A")
               ,   "COLUMN_B" = UPPER ("COLUMN_B")
               ,   "COLUMN_C" = UPPER ("COLUMN_C");
    {code} where every line above corresponds to a row fom the query.  The first line for every table will start with
    {code}
    UPDATE "TABLE_1" SET "but all the others will start with
    {code}
    and only on the last column for a given table will the statement be executed.                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

  • In ALV report is possible to display the subtotals in separate column

    Hi,
    I have made one ALV report, i need to show the subtotals of ekpo-brtwr (Gross Value)  grouped by or sorted by
    ekko-ebeln (Purshasing Doc) but in separate column  not at the end of the line items of each purshasing doc
    any idea how can i do this?
    let us say i have the following cols in my report:
    ekko-bukrs
    ekko-ebeln,
    ekko-aedat,
    kko-lifnr,
    ekpo-brtwr
    i want to add one more col for example (T-Gross) and to fill it with the suptotals of (ekpo-brtwr) sorted by the
    ekko-ebeln (Purshasing Doc).
    thanks a lot.

    If you need one more column, then you need to expand the field catalog table and data table as you did for all other fields and you need to code the logic to calculate the subtotals and populate the new column, but if you want to have row, that will be easy. For more info you can [refer |http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/4063c42a-2549-2c10-fab6-a8c99d98b9b5?quicklink=index&overridelayout=true]this:

  • GR and IR as separate columns in the BEx report

    Hi,
    I am working on a Purchasing report in which the user needs the snap shot of PO history,Like PO no.,Item,Netprice,PO Qty,Net order value,GR no.,IR number etc.
    I created a view on EKPO and EKBE.Here,based on the BELNR(Material document) field we are taking GR and IR values.(When VGABE =1 BELNR=GR and VGABE = 2 BELNR=IR).I created DSO to load the data.The problem is they need the GR and IR as separate columns in the BEx report.Could u please help me how to achieve this in BI?
    Thanks & Regards,
    Dinakar.

    Hi Neel,
    I need the BELNR has to be split into GR No.and IR No. separately based on VGABE field.
    Thanks & Regards,
    Dinakar.

  • Download int table into csv file with each column in separate column in csv

    Hi All,
    I want to download the data in internal table to CSV file. but each column in the table should come as separate column in csv format.
      CALL FUNCTION 'GUI_DOWNLOAD'
        EXPORTING
          FILENAME                = GD_FILE
          FILETYPE                = 'ASC'
          WRITE_FIELD_SEPARATOR   = 'X'
        tables
          DATA_TAB                = I_LINES_NEW
        EXCEPTIONS
          FILE_OPEN_ERROR         = 1
          FILE_READ_ERROR         = 2
          NO_BATCH                = 3
          GUI_REFUSE_FILETRANSFER = 4
          INVALID_TYPE            = 5
          NO_AUTHORITY            = 6
          UNKNOWN_ERROR           = 7
          BAD_DATA_FORMAT         = 8
          HEADER_NOT_ALLOWED      = 9
          SEPARATOR_NOT_ALLOWED   = 10
          HEADER_TOO_LONG         = 11
          UNKNOWN_DP_ERROR        = 12
          ACCESS_DENIED           = 13
          DP_OUT_OF_MEMORY        = 14
          DISK_FULL               = 15
          DP_TIMEOUT              = 16
          OTHERS                  = 17.
      IF SY-SUBRC NE 0.
        WRITE: 'Error ', SY-SUBRC, 'returned from GUI_DOWNLOAD SAP OUTBOUND'.
        SKIP.
      ENDIF.
    with the above values passd , I am getting csv file but all the columns in one column separated by some square symbol.
    How to separate them into different columns.
    Thanks in advance
    rgds,
    Madhuri

    Below example might help you understand on dowloading CSV file:
    TYPE-POOLS: truxs.
    DATA: i_t001 TYPE STANDARD TABLE OF t001,
          i_data TYPE truxs_t_text_data.
    SELECT * FROM t001 INTO TABLE i_t001 UP TO 20 ROWS.
    CALL FUNCTION 'SAP_CONVERT_TO_TEX_FORMAT'
      EXPORTING
        i_field_seperator          = ','
    *   I_LINE_HEADER              =
    *   I_FILENAME                 =
    *   I_APPL_KEEP                = ' '
      TABLES
        i_tab_sap_data             = i_t001
    CHANGING
       i_tab_converted_data       = i_data
    EXCEPTIONS
       conversion_failed          = 1
       OTHERS                     = 2.
    IF sy-subrc <> 0.
      MESSAGE ID sy-msgid TYPE sy-msgty NUMBER sy-msgno
              WITH sy-msgv1 sy-msgv2 sy-msgv3 sy-msgv4.
    ENDIF.
    DATA: file TYPE string VALUE 'C:\testing.csv'.
    CALL METHOD cl_gui_frontend_services=>gui_download
      EXPORTING
        filename                = file
      CHANGING
        data_tab                = i_data[]
      EXCEPTIONS
        file_write_error        = 1
        no_batch                = 2
        gui_refuse_filetransfer = 3
        invalid_type            = 4
        no_authority            = 5
        unknown_error           = 6
        header_not_allowed      = 7
        separator_not_allowed   = 8
        filesize_not_allowed    = 9
        header_too_long         = 10
        dp_error_create         = 11
        dp_error_send           = 12
        dp_error_write          = 13
        unknown_dp_error        = 14
        access_denied           = 15
        dp_out_of_memory        = 16
        disk_full               = 17
        dp_timeout              = 18
        file_not_found          = 19
        dataprovider_exception  = 20
        control_flush_error     = 21
        not_supported_by_gui    = 22
        error_no_gui            = 23
        OTHERS                  = 24.
    IF sy-subrc <> 0.
      MESSAGE ID sy-msgid TYPE sy-msgty NUMBER sy-msgno
                 WITH sy-msgv1 sy-msgv2 sy-msgv3 sy-msgv4.
    ENDIF.
    Regards
    Eswar

  • I have read 118 files (from a directory) each with 2088 data and put them into a 2 D array with 2cols and 246384row, I want to have aeach file on a separate columns with 2088 rows

    I have read 118 files from a directory using the list.vi. Each file has 2 cols with 2088rows. Now I have the data in a 2 D array with 2cols and 246384rows (118files * 2088rows). However I want to put each file in the same array but each file in separate columns. then I would have 236 columns (2cols for each of the 118 files) by 2088 rows. thank you very much in advance.

    Hiya,
    here's a couple of .vi's that might help out. I've taken a minimum manipulation approach by using replace array subset, and I'm not bothering to strip out the 1D array before inserting it. Instead I flip the array of filenames, and from this fill in the end 2-D array from the right, overwriting the column that will eventually become the "X" data values (same for each file) and appear on the right.
    The second .vi is a sub.vi I posted elsewhere on the discussion group to get the number of lines in a file. If you're sure that you know the number of data points is always going to be 2088, then replace this sub vi with a constant to speed up the program. (by about 2 seconds!)
    I've also updated the .vi to work as a sub.vi if you wa
    nt it to, complete with error handling.
    Hope this helps.
    S.
    // it takes almost no time to rate an answer
    Attachments:
    read_files_updated.vi ‏81 KB
    Num_Lines_in_File.vi ‏62 KB

  • How do I rank the KF Net Sales in a separate column?

    Hi all,
    I need to show ranks for Net Sales in a separate column.
    In one of my previous threads I received the reply that to get the rank in a new column I need to create a new CKF Net Sales Rank and refer Net Sales in this new Net Sales Rank and make the new CKF Net Sales Rank as (Calculate -> Single Values As -> "Ranked List (Olympic)").
    But when I do so I am getting dollar values of Net Sales in the new CKF "Net Sales Rank" and I do not want to see dollar values but what I want to see is the ranks like 1, 2, 3 based on the KF Net Sales column. The Net Sales rank column should be ranked based on the column Net Sales column.
    Please advise how will I get the ranks in the CKF Net Sales Rank column.
    Regards,
    TR.

    Hi TR,
    Once you do the calculate single values as Ranked List, you should get numbers, not $ values...

  • Select query output as a separate column

    Hi,
    How to represent the different values in the same table column as an output in the separate columns using the select query ?
    For example,
    Table "A" has column "col1"
    col1 contains values as below:
    col1
    ====
    1
    2
    3
    4
    5
    now want to display the above column values as a separate column as the output.
    col_alias1 col_alias2 col_alias3 col_alias4 col_alias5
    1 2 3 4 5
    How it can be done ?
    Regards
    Edited by: user640001 on Jan 31, 2011 11:19 PM

    Hi,
    You can try something mentioned in this link.
    http://www.oracle-base.com/articles/misc/StringAggregationTechniques.php
    or try this
    with t as (
    select 'PID12' as product_id,  1 as status_id, 'Time1' as time from dual union all
    select 'PID13', 2, 'Time2' as time from dual union all
    select 'PID14', 3, 'Time3' as time from dual)
    select min(case when status_id = 1 then product_id || ' ' || time end) as Recieved, min(case when status_id = 2 then product_id || ' ' || time end) as Accepted,
           min(case when status_id = 3 then product_id || ' ' || time end) as Delivered from tcheers
    VT
    Edited by: VT on Feb 1, 2011 5:28 PM

  • Need Parked amounts in a separate column in reports

    Hai Experts,
    I need your help. This is my requirement “Need Parked amounts in a separate column in reports”.
    How to deal with this issue?
    Could you please tell me step by step solved this solution.
    Advanced thanks,
    Regards,
    Sathis.J

    Hi,
    as you don't have a separate keyfigure for this you need to create a restricted keyfigure. Identify the characteristic value that is relevant for a parked amount and restrict your amount by this value.
    kind regards
    Siggi

Maybe you are looking for

  • HELP! itunes...i need help!

    OK. New to this chat and much of the apple stuff so please be patient. I have been using different ipods and itunes for years. got a nano few years ago and built up a library on an hp laptop. i got a couple giftcards in november/december of 2008. i u

  • Problem exporting '.txt' file size 23 KB and '.zip' file size 4 MB

    I am using Apex 3.0 version screen to upload '.txt' file and '.zip' file containing images. I can successfully export '.txt' file and '.zip' file containing images as long as '.txt' file size is < 23 KB and '.zip' file size < 4 MB from database table

  • IS it possible to speed up (fast forward) motion in imovie?

    IS it possible to speed up (fast forward) motion in imovie?

  • Note field in MIR7 Transaction

    Dear Friends, I need to get the Note field of MIR7 transaction in an ABAP report. Please suggest how can I get this data. Krishan

  • Outlook Express mail vanishes

    My email is supplied through Outlook Express. All was fine this morning. I opened up to check my email later on and found a totally blank page except for the default categories found on a new computer. Has anyone else had this problem? Does anyone ha